Editor’s note: Today’s post is by Jonathan Woahn, Co-founder of Cashmere.io, which helps publishers safely and responsibly monetize their content with artificial intelligence. Reviewer credit to Chef Tim Vines.

In Part One, The Missed Expectation, we argued that AI training will not become the durable revenue engine publishers hope for. While uncomfortable, this conclusion isn’t pessimistic. The real economic opportunity lies not in how models are trained, but in how they are used.

In this context, usage is the inference stage: the moment an AI model applies its training to solve a specific user problem. Clients pay to have these problems solved, not for the training data itself.

But inference monetization doesn’t arrive by default. Unlike the lump-sum payouts and corpus delivery for training deals, inference happens when the AI accesses and processes individual articles; capturing this usage requires new infrastructure and norms. Monetizing inference requires a clear understanding of how value is created and captured when content no longer reaches users in its original form.

In this second part, we focus on the path forward. Not a theoretical one, but one grounded in historical precedent and emerging reality. AI is not destroying the content economy; it is forcing it to evolve. To understand that evolution, we must first look to the past.

Robot hand and saving coins in glass jar. Artificial intelligence. Vector illustration

The Great Reallocation

When technology disrupts an industry, the way the product is consumed and the way consumption is paid for fall out of sync. Demand may even rise while revenue from established payment routes collapses. Only later do new revenue models allow producers to recapture value.

Music is the canonical example. Napster greatly expanded listening, but demolished revenue from CDs and records. So even though musicians continued making music, and users listened more than ever, the industry was unable to monetize this new form of consumption. Finally, after a period of turbulence, services like Spotify and Apple Music emerged with models that reconnected listening and compensation under a new consumption paradigm.

This pattern now applies to scholarly and professional content. Demand for authoritative knowledge has not declined: researchers, students, and professionals still rely on trusted sources to do their work.

What has changed is how that knowledge is delivered. Instead of directly reading articles, users increasingly ask AI systems to locate, synthesize, and report back on relevant work. Like the music industry in the Napster era, the delivery system has changed, but the revenue models are yet to stabilize.

Publishing is not at the end of its economic arc; it is in the turbulent middle of a reallocation.

Reallocations are often framed as cataclysmic and are very uncomfortable for those inside the affected industry. In practice, they are remarkably consistent. The table below shows how this reallocation unfolded in music and online video — and how it is now unfolding in publishing.

Table 1: The Reallocation Pattern

Stage Music Industry YouTube / Online Video AI & Publishing
1. Initial Shock (Demand Shifts) 1999 – Napster enables instant, free music sharing; listening explodes 2005 – YouTube makes video frictionless and global 2022 – ChatGPT introduces conversational, personalized answers
2. Monetization Collapse 2001-2002 – CD sales fall sharply; consumption decouples from payment 2005-2007 – Broadcast and DVD economics break; attention shifts online 2022-2024 – Pageviews and clicks lose relevance as answers bypass human reading
3. Existential Panic 2000–2001 – Napster lawsuits; industry fears extinction 2007 – Viacom lawsuit; “YouTube will destroy media” 2023–2024 – Publisher, author, and news lawsuits against AI labs
4. Token Partnerships (Unscalable Deals) 2003 – Labels license catalogs to iTunes (bespoke, fragile) 2006–2008 – Media companies strike custom YouTube deals 2023–2025 – Strategic AI licensing deals with select publishers
5. Infrastructure Formation ← We are here 2003–2008 – Licensing frameworks, DRM, billing systems 2007 – Content ID, Partner Program, attribution tooling 2024–2026 – Usage tracking, MCP, publisher programs
6. Sustainable Monetization 2008–2015 – Downloads + streaming + touring + merch 2010–2018 – Ads + subscriptions + creator ecosystems TBD – Subscriptions, ads, various licensing models (?)
7. Norms Lock In (Defaults Harden) ~2019 – Streaming normalized; revenue surpasses pre-Napster peak ~2018–2020 – Creator economy stabilizes; rules widely accepted TBD

This Reallocation Is Different

In every prior reallocation, the decisive moment came before monetization stabilized — when infrastructure was beginning to form, but norms had not yet locked in. What followed depended on whether industries correctly understood the shift they were actually facing.

AI places publishing at that same inflection point, and where we go from here will depend on how well we understand the shift we are facing. This time, it is not merely one of distribution.

Previous reallocations changed how content was packaged and distributed, but the content itself still reached the end user largely as the creator intended. A song moved from vinyl to CD to streaming, but listeners still heard the original recording. A film moved from theaters to DVDs to online video, but viewers still watched the work as originally produced.

AI breaks that continuity.

Content can now create value without ever being presented to the user in its original form. Ideas are extracted, combined, and reassembled into new outputs. Users benefit from the underlying works without directly encountering them. When value is created this way, revenue from that value does not automatically flow back to its sources.

History offers a playbook for how platform transitions unfold. What it does not offer are solutions for the latest reallocation, the route to reconnecting how product is paid for with how it is consumed. That is the problem the next section addresses.

Three Leading Economic Models Emerging for Publishers

If history is a guide, the market will not converge on a single economic model. Instead, it will stabilize around a monetization mix — coexisting approaches aligned to different sources of demand and control.

That mix is already forming in AI-driven inference. While implementations vary, three approaches are now emerging with clarity.

1. Pay-Per-Use (PPU)

The PPU model itself is straightforward. When a user consumes content through an AI platform, the platform pays the relevant publisher for the content that informed the response.

Crucially, the platform is not the economic consumer of the content. Users are. Through their queries, engagement, subscriptions, or attention, users signal which sources matter. The platform aggregates those signals and compensates publishers accordingly.

This is how Perplexity’s Premium Data for All program operates. When licensed publisher content is used to generate an answer, Perplexity compensates the publisher on a per-use basis. The license is narrow and temporary, but it is paid — and it scales with demand. Better sources improve answer quality, which drives usage and, in turn, supports the subscriptions and upgrades that fund content access.

The same pay-per-use logic can be funded by advertisers rather than users. ProRata.ai offers a clear example. ProRata applies an ad-supported model to AI inference: ads are displayed contextually alongside AI-generated answers, generating revenue on every query. That revenue is split 50:50 with publishers whose content contributed to the response. Users receive free access, advertisers fund the system, and publisher compensation remains tied directly to per-query usage.

In both cases, the unit of value is the same: content contribution at inference time. What differs is the source of funding.

While the economics are simple, the infrastructure is not. Attribution, contribution weighting, fraud resistance, and transparent reporting must operate reliably at query-level granularity. The closest analogue is pay-per-click advertising, as implemented by Google. As with PPC, Pay-Per-Use can scale quickly — but only once pricing logic, attribution, delivery, and enforcement infrastructure are in place.

2. Bring-Your-Own-License (BYOL)

BYOL models work differently. Here, the user already holds rights to the content — typically through an institutional or professional subscription — and chooses to connect those rights to an AI platform.

A clear example is Wiley’s partnership with Perplexity. Institutions that already license Wiley content can authenticate within Perplexity and access the material their existing agreements permit. In this arrangement, the AI platform acts as an interface layer, not a reseller, and the commercial relationship between publisher and customer remains unchanged.

The defining feature of BYOL is portability. The publisher sells the license to the customer; the customer decides where to exercise those rights. Like Netflix, a subscription is not tied to a single interface or device. BYOL extends that portability into AI-driven workflows, preserving pricing, trust, and institutional relationships while making platforms more useful to licensed users.

The constraint is reach. BYOL inherits the limits of existing entitlements, which means the completeness of AI-generated answers may vary across users or institutions depending on what content their subscriptions allow the system to access. In research settings, this creates a real risk: relevant work can be silently excluded if it falls outside the licensed corpus.

For that reason, BYOL works best as a portability layer for known, licensed collections — not as a complete solution for discovery when access must be created at the moment of need. That gap is what the next model addresses.

3. Licensing on Demand (LOD)

The third model centers on unlocking content at the moment it is discovered. While specific implementations are still emerging, several major publishers are actively developing Licensing on Demand offerings and are expected to bring them to market soon.

In Licensing on Demand, users do not need to hold licenses in advance, as with BYOL, nor rely on their AI platform to have pre-negotiated access, as with PPU. When valuable content surfaces during an AI interaction, access can be licensed immediately, in context, and under terms set directly by the publisher.

The core shift is timing: discovery comes first, and licensing follows immediately thereafter. Instead of forcing users to leave the AI experience or guess which subscription might apply, access is created precisely when the need becomes clear.

If Pay-Per-Use monetizes what users already consume, and BYOL extends licenses users already hold, Licensing on Demand enables access exactly when users — via AI-powered discovery — realize they need it.

Norms Worth Locking In Early

The AI reallocation right now offers publishers a rare opportunity: a chance to set the rules before they harden.

For the last two decades, search engines like Google and marketplaces like Amazon established defaults that publishers largely had to accept after the fact. AI is different. The ecosystem is still forming, and the terms are not yet fixed. That creates an opening to agree on how value should move.

From the publisher viewpoint, these norms should be stated early and often:

Inference on premium content should be paid.
If authoritative content improves AI’s answers, the publisher should be compensated when the content is used.

Attribution should be standard.
Source visibility benefits users, preserves trust, and reinforces publisher value.

Usage should be transparent.
Markets work when participants can see what is happening. Measurement enables pricing, investment, and accountability.

Direct relationships should remain intact.
AI can intermediate access without permanently distancing publishers from their audiences or customers.

These norms are straightforward. They align incentives, support sustainable markets, and reflect how publishers already operate elsewhere.

AI systems depend on high-quality content, and publishers still control the creation of that content. The earlier these expectations are reinforced across the industry, the more likely they are to become the defaults that shape what comes next.

Conclusion

AI is not going to pay for content the way many publishers once hoped. Training AI models will not become a broad, recurring revenue engine. AI will pay for the information it needs when conducting inference — when it accesses high quality information and brings that value to the end user.

The publishers who engage now — by aligning on shared expectations, supporting emerging models, and investing in common infrastructure — will help shape how knowledge is discovered, credited, and paid for in the AI era.

Jonathan Woahn

Jonathan Woahn

Jonathan Woahn is a co-founder of Cashmere, a platform designed to help publishers safely and responsibly monetize their content in AI-powered applications. He believes human-created content is what connects us and advances shared knowledge, and that creators and publishers need clear incentives to continue producing it. His work focuses on building systems that preserve compensation, credit, and control over how content is used, while enabling AI to responsibly incorporate high-quality, human-generated material at scale.

Discussion

21 Thoughts on "Guest Post — AI Isn’t Going to Pay for Content … Part Two: The Path Forward"

Jonathan: This is one of the most clear and insightful descriptions of productive ways in which to think about and act on AI’s influence on the future (and the now) of publishing. Thank you for taking the time to share your knowledge and expertise and for inspiring and energizing us.

Barbara Kline Pope, Executive Director, Johns Hopkins University Press

Thank you Barbara—for what it’s worth, I believe solving this problem is critical for not just the publishing industry, but also for AI. I don’t believe either can survive without the other.

Therefore, we need to figure out the right paths forward that allow everyone to thrive.

Thank you for the explanation of the three emerging economic models. Do exclusive partnerships ever become part of the discussion?

They’re part of the conversation right now—what I called “Token Partnerships” in the Reallocation pattern.

Ultimately that will be up to the publishers and if they want to go exclusive with different platforms. There’s also a model where the license could shift. I think about the TV show Seinfeld. Sometimes Netflix holds the license, sometimes CBS holds the license, sometimes Hulu (held) the license. It’s going to depend. There will be times a publisher has a well known, respected, desirable dataset, and they’re able to work out a preferable licensing arrangement with a lab to leverage that data and make it accessible to their users.

In my perspective, given the way user preferences seem to be orienting individuals with different models / ecosystems, I personally believe it’s going to largely be an exposure game—trying to meet users wherever they are. I’m not yet convinced that the platforms are as interchangeable as Netflix and DisneyPlus are.

This is a super helpful map of the “turbulent middle” we are in. I particularly appreciate the way you separate PPU, BYOL, and Licensing on Demand, and tie them back to concrete historical precedents in music and video. The emphasis on inference-level attribution, transparency, and preserving direct relationships feels spot on. From where I sit, that mix of economic models plus shared norms is exactly what we need if we want “knowledge objects” and Omnipub-style structures to underpin sustainable AI ecosystems rather than be strip-mined by them!

Thanks for this really excellent, clear eyed, and well written overview. And just to add: we need standards for usage and attribution! NISO and COUNTER Metrics are thinking a lot about this & it would be great to have your thoughts on what this could look like as something that happens alongside the monetized licensing pathways.

You are absolutely correct Monica. I’m working with the NISO team on this as we speak (I’m planning to be at NISO Plus next month, if you’re there let’s chat!), and we’re in contact with the Counter team on the same topic.

We’d love to be involved at the ground level to make sure the industry gets all the information needed to keep things moving.

This is great to hear- I won’t be at NISO Plus but both COUNTER and NISO leadership will be in the room. So glad you’re engaging on the metrics as well as the money.

Wonderful piece – and, thankfully, clear and concise providing excellent insight. The Napster analogy is very good. When books came online, effectively with companies like ebrary, Netlibrary, EBL, etc., the new technology presented opportunities to deliver content in new ways, such as the large subscription model (ebrary – then acquired by ProQuest, and shazaam, the ‘new’ PQ model now, but propelled by big-data brother Clarivate!). The new technology supported new models such as Demand-Driven Acquisitions, Short-Term Loan, and other disruptive models – phase 1 of the disruption. Book Wars (John Thompson) points to phase 2. And this article fleshes out where this is going for scholarly books in institutional markets. The days of the old fashioned booksellers, like the buggy whip makers before them, are numbered. But the outlook is sunny rather than drab and threatening!

I believe it is sunny! The challenge is, there will be casualties along the way—it’s inevitable, as not everyone has the means or manner to weather the turbulent times ahead. But those that do I believe will come out much stronger on the other side as a result.

Thanks Jonathon – on usage data and human user behavior, I believe there are generally 3 buckets of current platform usage – originally presented by ReadCube some 5-6 years ago
1) scan read 20-30 seconds – title, author, abstract digested – discovery
2) slightly more in-depth review – 2 minutes+ – check paper findings, citations etc
3) deep read 20+ minutes, digestion of full article content – aligned to researchers topical interests

This usage was generally split 1/3, 1/3, 1/3 across scholarly content and users reading behavior

Now with AI there’s a lot quicker assimilation, translation and summary of knowledge – and as you note, more interactive q&a with systems like ChatGTP around the knowledge and questions researchers have.

Is this changing in human usage behavior and how monetary value is going to be set by AI going to change ? And if so what should publishers be doing – the Music analogy users can hear a 20-30 second play of music but then have a choice to buy or know the music they want is part of their streaming service. How do we know with AI usage how scholarly content will be truly digested and used – and do publishers and AI need to know this?

It’s certainly a changing world – are OA models better in this world to get money up front from authors – or subscription paywalls ensure better payout from AI usage – all to be determined – good discussion, TSK article and future conference topic!

This is a great question Adrian—what should publishers be doing? I’d be lying if I said I _knew_ the answer. But I do have some informed hypothesis that we’re doing everything we can to act on.

What we know is the way content is consumed and monetized is going to look quite different from how it has looked traditionally. Context is everything for the AI, and the more _relevant_ context we can provide, the better it is able to help us. As “AI slop” continues to proliferate, the value of content created by Humans I believe will rise.

So then what’s needed is the infrastructure that allows content to be composable, transportable, and flexible to adjust to the rapid ways that content consumption is changing with AI. That will provide the creators with the flexibility to not have to buy into the “final” single solution today, but they can adapt as the market changes.

And agreed, this conversation is just getting started!

Thank you for these well-written pieces! They helped to open my eyes as to where things stand in this interesting crossroads of technology. I particularly appreciated the comparisons between other technologies that have disrupted the status quo. It appears that with the examples you gave, standards didn’t really get solidifed for 10-15 years after the “initial shock” of each industry. Do you think it will take that long with AI content or will those standards need to be solidified faster as they impact more day-to-day operations across so many use cases?

It’s hard to say, but the trends suggest everything is accelerating. I don’t think it’ll take 10-15 years, but it also isn’t going to happen all in this year (2026). It’ll take some time, but who knows…there’s a saying that says we tend to overestimate what will happen in the near term, but underestimate what will happen in the long term.

I feel that saying is even more true when it comes to AI.

Hi Jonathon. Thanks for iterating the path forward on how AI and scholarly communications may evolve.

I am curious about your analogies and it made me wonder if perhaps, scholarly communications publishers with powerful AI inference products will eventually explore syndicated delivery of same – either via global libraries consortia – or a new type of media? E.g. a health library consortia in say, Australia, where I am, purchases bundles through Wiley – and that comes with suitably proven and trustworthy AI inference tools behind the bundle etc. The agreement allows this health library consortia to redistribute/syndicate the tools across other types of special libraries – say, in universities – that have medical and nursing education programs – so that their use is built into medical training and thus facilitates how information is used and research is built upon in medicine, nursing and allied health in our hospitals.

Does that sound like a Thing? Or am I wish-casting onto AI inference tools something that publishers would be appalled by? Like how would they get the stats for that? And ergo, their attribution and payment?

Anyway… you just got me thinking and imagining what might be.

Hi Michelle,

Great comment, and I can respond from the Wiley perspective . The scenario you describe is definitely in line with the direction Wiley is going — we have developed much of the technical infrastructure to make this possible already, and we are now refining functionality and implementing business models for RAG access with corporate R&D subscribers. Wiley is running a few pilots with university libraries to test similar arrangements, and though it’s not my side of the business, I’m reasonably confident that if we see demand from institutions for this kind of arrangement, it will be delivered.

Still early days, so I think we’re a few years off from the syndication scenario, but it stands to reason that we’ll get there once we’ve hammered out the licensing, legal, and revenue models in a more standardized way.

I think it’s safe to say, if there’s a viable path to monetization that satiates concerns of the involved stakeholders, a path will emerge.

As Michael from Wiley notes in his comment above, many publishers are already headed in this direction—but it’s going to take some time for the markets to form. There’s a lot of foundation work that’s needed to unlock syndicated delivery.

I think the predictions here are spot on, but the optimism is falsely placed. Spotify and YouTube have certainly been a mixed bag in terms of financial outcomes for artists, while entrenching the monopolisation of industries.
Consider this reporting from CBC Canada:

‘”The numbers are wild — 1,500 artists made over $1 million from Spotify in 2024, and 100,000 artists generated at least $6,000. That sounds great, but when you realize that there are over 12 million uploaders, the competition is staggering,” Alper said.

“The vast majority of artists are still making pennies per stream, and unless you’re in the top few per cent of streamers, you’re probably not quitting your day job anytime soon,” he added.

Alper explained that major-label artists with massive streaming numbers can make substantial money, but for mid-level and emerging artists, streaming income is often unsustainable.’

Massification comes with other threats, too, and that is equally true in scholarly publishing’s relationship with genAI. If authors are reimbursed in this fashion, the trend will be an increase in volume and decrease in depth and quality, in order to capture payments. Bad for research, and bad for equity, but likely outcomes nonetheless.

Wayne,

I think it depends on how you define optimism. Based on your comments of Spotify, it seems like you’re suggesting “optimism” means more equal opportunity for everyone—and at no point in the article do I suggest this will be the case.

For me, the optimistic bet I’m making is that premium content publishing is going to become even more valuable than ever with AI, and the industry as a whole (academic, scholarly, trade, research, intelligence, media, news, journalism, etc) is going to grow. The unspoken implication of my bet is—growth doesn’t guarantee survival for everyone—and I think that’s possibly the type of optimism you may be referring to?

Before Napster, the music industry had the “Big Six” music labels (Sony, Universal, EMI, BMG, PolyGram, Warner). Today, there’s the “Big Three” (Sony, Universal, Warner). There was massive consolidation and acquisitions within the music industry; however, today in aggregate, it’s bigger than ever.

It’s very possible something similar could happen with publishing—especially as the lines between content providers, curators, and technology providers continues to blur.

Jonathan, this is a thoughtful breakdown of inference-based AI licensing economic models. I agree with your broad point that training model licensing is not a sustainable economic model (though it is delivering surprising lucrative short-term revenues for many organizations, and that is not a bad thing!).

I do have a quibble with your table. You are comparing the impact of LLMs on the publishing industry with the impact of the Web on the music and film/television industries. This skips the impact of the Web on the (scholarly) publishing industry and its economics models (which were profound). That “reallocation” was not trivial. Go ask the dozens of subscription agents that no longer exist or any librarian about The Big Deal. In that sense the table is apples and oranges. You are comparing the impact of one paradigm shift in computing on two industries, with the impact of another (potential) paradigm shift on another.

You are also positing a reallocation of revenues but one of your economic models (BYOL) would not reallocate revenues at all and it is not clear (as Wayne points out) that real money will materialize from either of the other two models (remember the “article economy”?).

Licensing has been part of the revenue mix for scholarly publishers for decades. Whether licensing to AI companies becomes the primary revenue stream for scholarly publishers (supplanting institutional licensing, which would be a great reallocation indeed) or an extension of existing licensing programs is an open question.

It is an important question (maybe the essential question for publishers). It is hard to know in advance of a large technology shift where the value will lie and how behavior will change. With the advent of the Web, many scholarly publications didn’t understand how many individual subscriptions would shift to institutional licenses and undervalued those licenses. Skating to where the puck is going to be sounds easy in theory but in practice it is much more difficult.

Michael,

I appreciate your thoughtful response, and I say “quibble away”! I love me some respectful disagreement and dialog. A few responses to your comment.

Regarding the table, I was trying to communicate what the reallocation pattern looks like using examples that 99% of readers would be able to identify with, even without a deep history of scholarly publishing. That said, I didn’t consider the example you shared—it was before my time—and you’re right, I can imagine it wasn’t a trivial shift.

I’m curious why you feel why the pattern is so apples and oranges? I recognize fundamentally the industries themselves are very different, but the pattern itself feels precisely like what’s happening today. For instance, ChatGPT has amped up buzz around bringing ads into their platform, which is a classic signal that we’re getting into stage 6. The playbook is/has been happening right around us.

Regarding the economic models and “skating to where the puck is going to be”—it’s hard to prognosticate, and candidly—puts one in quite a vulnerable position as anyone who disagrees with your view of the future will tell you why you’re wrong. My question in these situations is, can you tell me more? I’d love to understand better why you disagree and where _you_ think things are headed?

My intention of this article is to (a) provide insights into what is _actually_ happening in the market right now based on facts and (b) redirect attention to inference as the real opportunity for publishers, and to stop chasing the false dream of training licensing as the AI licensing gold mine.

Comments are closed.