Editor’s Note: Today’s post is by Mike Olson, Assistant Professor and Cataloging Librarian at Murphy Library, University of Wisconsin-La Crosse. Building on “Classification as Colonization,” this examination reveals how vendor consolidation and AI automation are systematically eliminating the invisible labor that sustains library systems.
The recent layoffs at OCLC, the Dublin, Ohio-based nonprofit that manages the Dewey Decimal System and WorldCat, offer a stark glimpse into the precarious future of library and information labor. In July 2025, OCLC confirmed it reduced its central Ohio workforce by about 80 positions, citing “shifts in technical skill requirements, growing influence of artificial intelligence, and ongoing changes in higher education and libraries.”
The timing reveals a fundamental contradiction in how cataloging expertise is valued. While OCLC cites AI as justification for workforce cuts, the organization has also announced it is developing AI tools for metadata processing. Here lies the paradox of our moment: the same technological advances celebrated for their efficiency are erasing the human expertise that creates the high-quality metadata these systems depend on to function.
This contradiction becomes intelligible within the broader transformation of universities from public good institutions to market-driven entities focused on knowledge commodification. Library labor exemplifies what Maurizio Lazzarato identifies as “immaterial labor” — work that produces both informational content and cultural standardization, becoming directly productive for digital capitalism through vendor-controlled systems that aggregate and monetize standardized metadata.

The Invisible Infrastructure Under Attack
The devaluation of cataloging expertise becomes visible when institutions eliminate these positions during budget cuts. Northwestern University’s announcement of layoffs affecting 425 positions included Violet Fox, one of the field’s leading voices in critical cataloging and creator of the Cataloging Lab. Fox’s dismissal is particularly telling — her work on reparative description practices exemplifies exactly the kind of specialized cataloging knowledge that institutions are eliminating precisely when it’s most needed. Catalogers like Fox perform the invisible work of challenging colonial classification systems, ensuring that subject headings reflect community needs rather than vendor-supplied generic terms, making collections discoverable and accessible, and maintaining the metadata quality that enables scholarly research.
From Card Catalogs to Vendor Control
The nostalgia for card catalogs that occasionally surfaces in library discourse isn’t mere romanticism; it reflects a genuine loss of institutional autonomy. The physical card catalog represented direct institutional control over bibliographic organization, where local catalogers made decisions about subject headings, classification, and cross-references based on their collections and communities’ needs.
Today’s “online catalogs” increasingly function as interfaces to vendor-controlled discovery systems that aggregate metadata from multiple sources with minimal local oversight. Collection Discovery Indexes (CDIs) like EBSCO’s Discovery Service or ProQuest’s Primo Central dump millions of records from contracted vendors directly into library search interfaces, creating what appears to be a comprehensive catalog but actually represents a fundamental transfer of intellectual control from local institutions to corporate entities.
This transfer represents platform capitalism’s colonization of library infrastructure. Where card catalogs embodied local cataloging decisions, contemporary discovery systems operate through what Shoshana Zuboff identifies as the extraction architectures that gather user behavioral data while providing search results optimized for vendor algorithms rather than local collection strengths or community needs.
The metadata quality crisis in vendor-supplied records typifies this loss of control. Local catalogers historically ensured consistency and accuracy through direct engagement with materials. Vendor-supplied metadata often contains errors and generic subject headings that reflect automated processing rather than intellectual analysis. Yet libraries have a limited ability to correct these records without disrupting system updates that overwrite local modifications.
Vendor Consolidation and Legal Battles
The OCLC layoffs occur within the broader vendor consolidation that has fundamentally altered power dynamics within library ecosystems. Multiple legal battles illuminate these tensions. OCLC’s recent settlement with Clarivate over the MetaDoor dispute forced Clarivate to cease development and permanently delete work product that incorporated WorldCat records. As The Scholarly Kitchen reported, this lawsuit revealed deep tensions over who controls bibliographic metadata created through collaborative library labor. Even more revealing is OCLC’s ongoing litigation against Anna’s Archive, accused of scraping WorldCat’s data.
What’s particularly striking is how OCLC’s litigation strategy positions libraries. While OCLC doesn’t typically sue libraries directly, its lawsuits create a chilling effect that constrains how libraries can use their own contributed metadata. Libraries find themselves contractually restricted from sharing WorldCat records outside OCLC-approved channels, sometimes unable to freely reuse catalog data they themselves created and contributed to the collective database.
This reveals a deeper tension: WorldCat emerges from decades of collective cataloging labor by thousands of librarians, yet OCLC asserts proprietary control over this collaboratively created resource. Libraries become trapped in a system where their intellectual labor is incorporated into a collective good, but access to that resource remains subject to corporate control and contractual restrictions.
AI’s False Promise: From Cataloging to Discovery
The promise of AI in library cataloging is seductive: faster processing, reduced manual entry, freeing workers for “more important” tasks. Yet the Library of Congress’s recent experiments found that Large Language Models scored only a 26% F1 when predicting Library of Congress Subject Headings, while subject classification models achieved only 35% accuracy. These results reveal a fundamental shift in cataloging labor from record creation to training, evaluating, and curating algorithmic outputs—catalogers now find themselves providing feedback and quality review for systems that cannot replicate the human expertise, cultural knowledge, and contextual understanding that quality metadata requires.
The disconnect between AI promises and library reality was already starkly apparent at the 2024 Ex Libris Users of North America (ELUNA) conference. While Ex Libris staff delivered polished presentations touting AI integration into every facet of their products, from customer service to cataloging, user sessions revealed a different story entirely. Librarians shared workarounds for basic functionality that had remained broken for years, struggling with vague documentation and systems that failed to deliver even simple results like getting resources to appear prominently in search results. The tension was palpable during Q&A sessions, where librarians voiced disinterest in AI features while grappling with fundamental system problems.
More troubling is how AI discovery interfaces bypass cataloging work entirely. Aaron Tay’s investigation exposed how Clarivate’s Summon Research Assistant and Primo Research Assistant refuse to search terms like “Tulsa race riot,” “Gaza War,” or anything containing “massacre,” returning zero results from billion-record databases. Content-filtering layers designed for social media chatbots now “quietly block controversial topics from being searched” in academic library systems, undermining the careful subject analysis that catalogers perform to ensure comprehensive access to scholarly materials.
This represents two simultaneous threats to library infrastructure: AI cataloging systems that require extensive human oversight while promising to eliminate catalogers, and AI discovery interfaces that replace structured metadata searching with chatbot interactions that can silently censor topics. Where traditional library catalogs surface resources through cataloger-assigned subject headings and classification numbers, AI discovery tools depend on algorithmic interpretation that can simply refuse to search—making catalogers’ intellectual work invisible while constraining access to the very materials they’ve carefully described and organized.
Implications for Scholarly Communication
The displacement of cataloging expertise carries profound consequences for how scholarly knowledge circulates. Publishers increasingly find themselves dependent on vendor algorithms rather than subject-specialist catalogers to connect their content with researchers. Where human catalogers historically ensured that resources could be discovered through multiple access points—alternative subject headings, cross-references, and classification schemes that reflect disciplinary nuances — AI discovery systems operate through black-box algorithmic interpretation that prioritizes vendor metrics over scholarly relevance.
This shift fundamentally alters the relationship between content creators and researchers. Vendor-supplied metadata often contains errors and generic subject headings that reduce precision in academic search results, making it harder for scholars to locate relevant sources and for publishers to demonstrate usage and impact. The transformation from locally-controlled catalogs built on human expertise to vendor-mediated platforms means that discoverability increasingly depends on corporate decisions about interface design, search algorithms, and content filtering rather than intellectual analysis of scholarly materials.
Most troubling is how AI implementation can actively constrain academic inquiry. When discovery systems can refuse to search entire categories of terms or topics, they undermine the foundational principle that researchers should have comprehensive access to relevant materials. As Jay Singley warns, “Uncritically adopting AI tools in discovery systems will perpetuate, if not exacerbate, existing biases and suppression of minoritized people”—a concern that extends beyond representation to fundamental questions about what knowledge remains discoverable in an algorithmic environment.
The Price of Automation
The crisis facing library labor today extends the colonial dynamics explored in “Classification as Colonization.” Just as classification systems can perpetuate epistemic violence through seemingly neutral organizational choices, labor policies can systematically devalue and eliminate the human expertise that makes libraries function.
The 80 workers laid off at OCLC, the elimination of Violet Fox’s position at Northwestern, and thousands of other displaced catalogers represent more than statistical casualties—they embody decades of accumulated expertise in organizing human knowledge that, once lost, cannot be easily recovered. When AI systems can silently block searches for “Gaza War” or “Tulsa race riot” while achieving only 26% accuracy on subject classification, we witness the replacement of human intellectual judgment with corporate algorithmic control that serves surveillance capitalism rather than scholarly inquiry.
Understanding this crisis requires theoretical frameworks that connect institutional transformation, technological change, and worker experience within broader patterns of academic capitalism and surveillance capitalism’s penetration of knowledge work. As libraries become more central to digital knowledge production and commodification, cataloging labor becomes simultaneously more visible, more technologically mediated, and more affected by commercial and algorithmic logics.
As Langdon Winner’s foundational question — “Do Artifacts Have Politics?” — reminds us, library technologies embody social relations and power structures rather than neutral tools. The challenge isn’t whether AI will change library work, it’s whether those changes will serve the public good or private profit, whether they’ll enhance human expertise or eliminate it, whether they’ll democratize knowledge or concentrate control in vendor platforms that can censor at will.
The stakes could not be higher. In an era of information abundance and epistemic chaos, libraries serve as crucial sites for democratic knowledge practices. The catalogers and information workers being eliminated today are the same workers who challenge colonial classification systems, create community-responsive metadata, and ensure that controversial but essential topics remain discoverable. Protecting these workers isn’t just about labor justice, it’s about preserving the infrastructure of informed citizenship itself. The question isn’t whether libraries will survive digital capitalism’s transformation, but whether they can become sites of resistance to it.
Discussion
14 Thoughts on "Guest Post — Beyond Classification: The Human Cost of Library and Information Labor Under Digital Capitalism"
This is such an excellent take and exceedingly well put. Thanks for writing this.
Thanks for this post. Coincidentally I just dived into https://cjal.ca/index.php/capal/issue/view/2425, about Refusing Crisis Narratives in Academic Librarianship. Worth reading again in the light of current developments in Librarianship!
Hi, Hubert —
Thanks so much for sharing that — Refusing Crisis Narratives in Academic Librarianship is a powerful and timely read. I agree, revisiting it in light of recent developments really sharpens its relevance. The pushback against crisis framing feels especially urgent now, as we see how narratives of inevitability around automation and restructuring can obscure the deeper systemic choices being made.
I appreciate the connection — it’s a great companion piece to the themes I was exploring.
If “the question isn’t whether libraries will survive digital capitalism’s transformation, but whether they can become sites of resistance to it.” The answer is obvious. Libraries throughout history and in all literate societies have been created and maintained by the powerful in that society. Libraries are expensive, delicate, and complex operations. They are difficult to create and to sustain over the long term. It takes power to marshal the resources for them. When power changes radically they tend to collapse (see Roman Empire in the west, regime and religious change in pre-modern India, dynastic change in China, Spanish colonization in the Americas.) Only to be rebuilt by the newly powerful. There are occasional “sites of resistance” but they tend to be small, ineffective at resisting the powerful, and (relatively) short lived. By joining with other (powerful) actors in society that seek to influence the direction of our economic system, how it is regulated, and how the surplus created by that system is distributed, we may influence it for the better, but I think the ‘sites of resistance’ paradigm is the road to defunding and irrelevance.
Hi, Jonathan —
Thank you for engaging so deeply with the piece and for bringing historical perspective into the conversation.
You’re absolutely right that libraries have often been sustained by dominant powers, and that their survival has historically depended on the stability and priorities of those powers. The examples you cite—Roman collapse, dynastic shifts, colonial disruptions—are powerful reminders of how fragile cultural institutions can be.
That said, the “sites of resistance” framing isn’t meant to deny those realities, but to ask what libraries might become within the contradictions of digital capitalism. Resistance here doesn’t necessarily mean overthrowing systems of power, but rather creating spaces—however modest—where alternative values can be practiced: equity, care, collective knowledge, and labor dignity. These efforts may be small, but they can still be meaningful, especially when they connect with broader movements for economic and social justice.
I agree that resistance without strategy or coalition risks marginalization. But I also worry that aligning too closely with dominant actors, especially in a system that commodifies knowledge and labor, can erode the very values libraries are meant to uphold. The challenge, as you suggest, is to navigate that tension: to remain viable while still advocating for a more just and inclusive knowledge ecosystem.
Thanks again for pushing the conversation forward. I’d welcome your thoughts on what kinds of coalitions or strategies you see as most promising.
As far as strategies for creating alternative spacies, I’ve read of a number of libraries that have chosen to end their affiliation with OCLC and go their own way. (Quite a number of European libraries have done this.) Closer to home, libraries whose collections don’t neatly square with typical cataloging would be candidates. LCSH is poor when it comes to drama, actual scripts of plays and I’m sure there’s a number of subject areas that are similar. There is archival description which remains a fairly local (or consortial) business unhampered by a central controller. So the bottom line is that the more unique and individualistic a collection, the more freedom it has to assert its own independence.
Bob, would you be able to direct me to libraries that have chosen to end their affiliation with OCLC? The economic reality is that we need to explore a way to do so.
Fascinating post and discussion.
The whole of Australia recently ended a longstanding arrangement whereby our national bibliographic database was provided to OCLC in exchange for access to WorldCat, inter-library loan software (VDX), and a few other tools. This arrangement ceased in July this year when OCLC failed to reach a mutually satisfactory renewal of the agreement with the National Library of Australia. Australian ILL is now managed through the Trove Partner Resource Sharing system, powered by ReShare (see https://www.library.gov.au/news-media/national-library-australia-set-evolve-national-resource-sharing). Of course, we have the advantage of having retained a national shared catalogue managed by the NLA.
Hi, Mike —
I have a question about this sentence:
This contradiction becomes intelligible within the broader transformation of universities from public good institutions to market-driven entities focused on knowledge commodification.
When you say that universities used to be “public good institutions,” do you mean to suggest that the university once met the economic definition of a “public good” (i.e. a non-rivalrous, non-excludable good), or do you mean that they once were, but are no longer, dedicated to promoting public welfare?
Hi, Rick —
Thanks for the close reading and for raising such an important question.
In that sentence, I’m using “public good institutions” in the broader sociopolitical sense — referring to the university’s historical role in promoting public welfare, civic engagement, and the democratization of knowledge. While the economic definition of a “public good” (non-rivalrous and non-excludable) is certainly relevant to discussions about education, my emphasis here is on the shift in institutional mission: from serving the public interest to operating under market logics that prioritize efficiency, competition, and monetization.
This transformation has profound implications for library labor, especially as we see expertise increasingly devalued in favor of automation and vendor-driven solutions. I appreciate your question — it helps clarify the stakes of that shift.
An interesting piece, thank you. I wanted to address your point on the guardrails in the Primo and Summon Research Assistants. We’ve been actively investigating the issue and wanted to share more context around what we’re doing to improve the experience.
Most of the blocking stemmed from Azure’s built-in content filtering, which is designed to prevent unsafe or illegal content. These filters are broad by design, and while they protect against genuinely harmful prompts, they can sometimes inadvertently limit access to legitimate academic topics. We’ve tested lower sensitivity settings for the Azure content filtering — moving from “medium” to “low” — and the results are very encouraging. Queries around historical or socially significant topics that were previously blocked are now returning results, while harmful content remains blocked. We have updated the setting for all Primo and Summon customers early this week.
While this won’t resolve every content filtering issue, it marks a significant step forward, and we ultimately believe that the solution lies in the collaboration between vendors, libraries, and researchers to define standards and advocate for provider-level flexibility that supports academic needs. We’ve made content filtering a key topic for our recent Academia AI Advisory Council and are exploring the formation of an Academic AI Working Group to help guide these efforts.
Hi, Amy —
Thank you for engaging with the piece and for providing updates on the Azure filtering adjustments within Primo and Summon. It’s encouraging to hear that sensitivity settings have been revisited and that more historically and socially significant queries are now returning results.
Still, the deeper concern persists: when discovery tools rely on opaque, third-party filtering systems—especially those designed for commercial or general-use contexts—library workers and researchers are left navigating constraints that are neither transparent nor accountable to scholarly values. Even with improvements, the fact that legitimate academic content was previously blocked underscores the precarity of outsourcing core discovery infrastructure to systems governed by corporate risk models.
This shift also sidelines the expertise of library workers, particularly catalogers, whose labor has long ensured that bibliographic records are rich, reliable, and rooted in disciplinary knowledge. Instead of surfacing sources based on that metadata, these AI assistants often deliver filtered summaries shaped by external moderation systems. The result is not just a technical workaround but a fundamental reorientation of discovery—from structured, intentional description to algorithmically mediated interpretation.
I appreciate the mention of collaborative efforts like the Academia AI Advisory Council and the proposed Academic AI Working Group. I hope these spaces will center the voices of library workers and researchers who experience the downstream effects of these decisions daily. Ultimately, the goal should be not just technical fixes, but structural shifts that prioritize academic freedom, labor equity, and epistemic justice in the design and governance of AI-powered tools.
Thanks again for contributing to the conversation.
Can you expand a bit more on what the algorithm considers “harmful”? If something is searching Primo, I expect it to be able to return any item cataloged in Primo. Plenty of researchers study harmful things — Nazi propaganda to pick an obvious example — and need those primary sources, regardless of how disturbing and inappropriate they would be in most other contexts. I want an algorithm that returns things relevant to my search, not that someone has decided are good for me.
I too am wondering what would be considered “harmful” in this scenario. I can understand the need for content filters on some AI chatbots — I am thinking particularly of the recent news story about Adam Raine, a teenager who was encouraged by ChatGPT to hide his suicidal tendencies from his parents. But that’s not really the sort of use case one would expect from your tools.
Further, it troubles me that a censored search would just return zero results, rather than informing the user that Primo/Summon is unable to complete their query. Users deserve to know that they are being denied access to information, rather than being met with an interface that suggests that info just isn’t there.