In November it will have been three years since ChatGPT was launched, and what a couple of years it has been. ChatGPT has changed the way we think about generative AI, both as a sector and as a society, with its accessible, conversational interface that enables anybody with a web browser to interact with it. In other words, it took the progress that computer scientists have been making for decades and made it directly visible to the public.

It’s no surprise then, that the Gartner hype cycle put generative AI right at the pinnacle of the peak of inflated expectations in 2023. As with any innovation, there was both excitement and skepticism in those early days. There were, and still are, strong concerns around attribution, the risk to copyright, hallucinations, low-quality content overload (aka slop); some even saw it as an existential threat to creativity itself. I acknowledge these risks, but I’m not overly concerned. Fear of risk shouldn’t prevent us from experimenting and innovating, instead they should be carefully monitored and mitigated. For example, privacy and security risks can be mitigated through approaches like guardrails, while risks posed by functional issues like hallucinations can be tackled by choosing the right technology for the right problem, for example retrieval augmented generation (RAG), which enables LLMs to incorporate new information in their answers without retraining as well as directing users to the sources on which claims are based.

From an innovation point of view, a more interesting question is whether we’ve really seen use cases that measure up to the huge amount of effort and expense that have gone into building the LLMs that power many generative AI applications. In a post I wrote about a year ago, I wondered out loud when the crash in the hype would come and, assuming things follow the hype cycle model, what would lie on the other side of the so-called trough of disillusionment.

While I’m still not sure when all this AI excitement will turn in on itself, or how deep the disillusionment will go, I find myself wondering whether, a year later, we know any more about what this technology will eventually do for us?

smartphone screen showing various AI apps: ChatGPT, DeepSeek, Anthropic Claude, Perplexity, Google Gemini, Microsoft Copilot.

How have things evolved over the last year or so?

Early predictions of the role of generative AI in academia and publishing focused on productivity, research, and writing support. Making writing easier and quicker is an obvious use case for something called a language model, but concerns have been raised about the risks to academic integrity and cheating, leading to what seems like something of a split in this part of the market. On one side, there are tools — which I won’t name here — marketed as being able to do your homework for you and not get caught by AI detectors. However, others have taken a more ethical approach. Grammarly and Writeful, who I mentioned in my post last year, have continued to develop responsible AI tools that help users improve their own writing, rather than doing it for them. We’re now starting to see tools that act more like AI-powered writing coaches that don’t necessarily generate your text directly, like Thesify, with its writing critiques and suggestions of what literature to read, and Profectus, which specializes in grant proposals and also offers services like training workshops on how to get funded.

Discovery and summarization tools were also among the first real product types to emerge, and now pretty much all of the web-scale discovery services have their own AI-powered natural language search and discovery interfaces. Meanwhile, startups like Elicit, ResearchRabbit, Scholarcy, Scite, Perplexity, and Proemial, among others, continue to make progress refining their target markets and use cases. One interesting area is AI summarization for specific audiences. Most of the examples I’ve seen so far are intended to make content more accessible for the general public, students, or researchers in an adjacent field. Solutions that are compellingly tailored to business applications seem to be still maturing. One example is Boardly, which is an early tool designed to help in the creation of board materials, and there are also products like Google’s NotebookLM, and QuillBot’s Executive Summary Generator. In the future, I think we’ll likely see more offerings like Iris and Questel, which integrate with external and internal documents and data, combining AI agents with technologies like RAG to help leaders make good strategic decisions without getting bogged down in information overload.

Automated Scholarly Paper Review (ASPR) is a whole new field that’s exploding the use of AI to protect research integrity. It includes tools for editorial triage, event management, formatting, scope checks, and reviewer matching. Clear Skies and Signals have received the most attention over the last year or so, with their network analysis approach to detecting patterns of research fraud. There are many fraud detection and integrity check workflow tools on the market, many of which focus not on network analysis, but on detection at the point of review. The sheer number of them may be evidence that the barrier to entry for building a workflow tool may not be very high. Examples include Morressier’s Effortless peer review, Editorial Pilot from Integra, Peer submit, and startups like Penenlope, reviewer zero, reviewer one, reviewer3 (is anybody noticing a pattern, here?), and newcomer nicia.

With so many options, it’s easy to feel overwhelmed, and with many of these tools being stand-alone products, they need to be integrated into existing workflows without unduly increasing the time needed and the cost of screening manuscripts. Several approaches have emerged to meet this challenge. The STM Integrity hub continues to bring tools together on a shared platform. Offerings like ReView 3.0 from River Valley Technologies provide integrated publishing workflow solutions, while Molecular Connections’ Research Integrity Toolkit focuses on integrating tools with existing platforms. Meanwhile some larger commercial publishers like Springer Nature and Wiley are partnering with technology companies to integrate ASPR directly into their own platforms. 

Where are things heading?

It’s clear that there is a lot of activity around AI innovation in publishing. The early trends in discovery, summarization, and workflows, particularly with editorial and research integrity, are still proliferating but are now showing signs of maturing and becoming more integrated. 

It’s still hard to see what will happen in the longer term, but one popular idea, as articulated recently by Eric Schmitt, ex-CEO of Google, is that user interfaces will go away, to be replaced by AI agents. What that looks like for scholarly publishing and for academia remains to be seen. Perhaps the trend towards expanding publisher services upstream will continue beyond author tools and we’ll eventually see AI agent-driven research and dissemination workflows all the way from the lab bench to the readers’ screen. Maybe the form of research publication itself will finally change and we’ll see agentic publications where a body of work, submitted as something like a research object, can be built on, interrogated, or synthesized along with other research through a natural language interface. Perhaps that will inevitably lead to the long-promised disruption to how research and researchers are evaluated and incentivized. Who knows? All I’m willing to say confidently at the moment is that the workflow and efficiency enhancements that AI currently provide are likely to continue to gain pace in the short to medium term and we still don’t know where this journey will eventually end.

Phill Jones

Phill Jones

Phill Jones is a co-founder of MoreBrains Consulting Cooperative. MoreBrains works in open science, research infrastructure and publishing. As part of the MoreBrains team, Phill supports a diverse range of clients from publishers and learned societies to institutions and funders, on a broad range of strategic and operational challenges. He's worked in a variety of senior and governance roles in editorial, outreach, scientometrics, product and technology at such places as JoVE, Digital Science, and Emerald. In a former life, he was a cross-disciplinary research scientist at the UK Atomic Energy Authority and Harvard Medical School.

Discussion

9 Thoughts on "Three Years After the Launch of ChatGPT, Do We Know Where This Is Heading?"

You said: “strong concerns around attribution, the risk to copyright, hallucinations, low-quality content overload (aka slop); some even saw it as an existential threat to creativity itself. I acknowledge these risks, but I’m not overly concerned.”

I am overly concerned! In a world of disinformation, misinformation and straight-up lies scholars need to be trustworthy and offer credible insights based on evidence and careful analysis. I think scholars need to draw evidence from others whose work they can study and if reliable, honor with a citation. I think we have a choice right now: stand on the shoulders of giants or stand on the backs of writers whose work was stolen.

What I’m observing is an increase in vocal AI resistance from academics, researchers, and readers. The vocabulary keeps expanding to include not only AI slop, but also terms that reflect larger societal, economic, and environmental issues such as “technofeudalism” and “digital colonialism.”

In the US resistance is heightened because we see AI tech bros chumming around with Trump, while noticeably silent when cuts are made to education, research, libraries, and museums. The message they give is: you don’t need to read, study, think, or create art, just let us take the reins. A lot of us are saying: no thanks!

Thank you for your comment. These are legitimate concerns, I completely agree. On the other hand, there’s a lot of ground being covered here and it’s important to break things down into specific issues. There’s no putting the AI genie back in the bottle, so we must figure out A) what’s coming and B) how do we adapt or cope.

Disinformation, misinformation, malinformation, and its variants represent a series of problems in multiple contexts. Industrial fraud and papermills are different to individual research fraud cases, which again is different to images posted on social media with misleading captions, which again is a different problem to bots flooding the social media zone with inaccurate memes that push an agenda. It’s a lot, I get it. On the other hand, the problem here isn’t technology itself but the misuse of it. That’s why I say that we should tackle these issues by putting guardrails in place and making sure the right tools are used for the right jobs.

By saying this, I don’t mean to dismiss the amount of work needed to make sure that AI is used in a safe and responsible way, but as the examples I pointed out show, people are actively working on these issues. For example, frameworks are being developed for the licensing of content for training AI or for inference, and regulation is starting to take shape with ISO/IEC 42001, the EU AI act and the UNESCO recommendations on the Ethics of AI. I didn’t go into a lot of detail about that side of things in my post only because it wasn’t the focus.

Important topic, but the coverage struck me as somewhat uncritical. “Concerns have been raised…” Right.

The only currency we have in scholarly publishing is the reader’s confidence.

AI-based “tools” are making it very difficult for me to be sure that I will be able to retain that confidence in any consistent way. We published this editorial on the topic this month, by coincidence, which shows one way in which we’re struggling to do that: https://journals.lww.com/clinorthop/fulltext/2025/10000/editorial__ai_assisted_letters_to_the_editor_scope.1.aspx.

But the wave of what my grandmother would have called “dreck” is cresting ever larger. We’re now seeing systematic reviews and meta-analyses that look to have been created using these tools almost completely. I don’t have the means to verify the references in articles of this sort, where the reference list for a single article may have > 100 sources; evidence—and my own experience with these—suggests that 30-40% of which are hallucinated. (See https://www.jmir.org/2024/1/e53164/, and https://link.springer.com/article/10.1007/s00146-025-02406-7).

I can see why authors would want to use these tools, but authors are not the most-important group to keep in mind. Readers—and, for medical journals, the patients whom those readers serve—are the group we need to privilege in our line of work.

How about some coverage of how we might do that more effectively?

Many thanks for considering.

Thanks for your comment.

I didn’t mean to be uncritical, it’s just not the area of the discussion I focused on. If you click on the tag for Artificial Intelligence at the top of the post, you’ll see quite a bit of coverage of AI from a variety of perspectives.

Thanks for those links and the suggestion for a post topic. From talking to research integrity folks, mostly through my work with STM, a lot of people are looking at automation and integrating tools that accelerate editorial checks into their existing workflows. In the post, I talked about a lot of the workflow and integrity tools that are being developed. Some of them are supposed to be able check references for validity and flag bad ones.

The next post I’m planning is on the basics of licensing content for AI use but perhaps I can look in more depth at the relationship between how research integrity tools (some of which are AI based themselves) may or may not make it easier to maintain the trust of readers.

The development and implementation of AI seems largely driven by the pursuit of power and profit, at the expense of human creators and consumers. As a librarian, that makes me extremely critical of hand-waving away legitimate concerns about AI just because it’s emerging, innovative, and trendy. While some AI tools are useful for specific tasks, that doesn’t negate the fact that our information landscape is turning to slop, and the tech companies have a profit incentive to make users rely on LLMs that they trust blindly instead of developing information literacy skills. Science and scholarship cannot thrive when the proliferation of AI means the reliability and credibility of the written word are impossible to judge.

I’m sorry if you think I’m waiving away concerns. That just wasn’t the focus of my post. If you click on the tag for Artificial Intelligence at the top of the post, you’ll find lots of posts about AI, many of which are critical and specifically about the issues you list in your comment.

That said, I’m not sure I agree that ‘our information landscape is turning to slop’. There are a lot of dedicated editors and reviewers that continue to distinguish between good research and poor quality content. Those publications continue to be aggregated and indexed and made available through web-scale discovery tools that serve specific academic communities. The task of filtering has become more difficult for editors, admittedly, but that’s not entirely the fault of AI. Just before the launch of ChatGPT, COPE and STM released a landmark report on the risk posed by paper mills to the scholarly literature, which demonstrates that pollution of the literature does not require LLMs. There are many initiatives, programmes and startups all working on research integrity challenges. At the risk of self-promotion, I’m personally working with STM alongside a number of other stakeholders on how research images might be cryptographically signed at source, as a means to safeguard against individual researcher fraud and AI fakery. So, these issue are taken seriously, it just wasn’t the focus of my post.

What has definitely changed is that searching google and google-like search engines no longer produces the quality of results that it used to, and social media is increasingly unreliable as a source of news and intelligent discussion. To be honest, that’s been a long time coming. What’s killing search engines isn’t AI on it’s own, it’s the business model of selling traffic, and the overuse of search engine optimisation by companies to draw users into low-effort content to sell stuff. That’s not an AI problem, it’s an enshittification problem, AI just helped it along.

Thank you for this post! I see AI less as the endgame and more as the beginning of something we have yet to imagine. We have become used to “new innovations” coming out at scheduled times, slight improvements to already established technology that has often been intentionally held back in order to meet the schedule. AI is a drastic departure from this. And while I do not know if some of its primary issues can be solved, this technology is the beginning of Something Else. While there are challenges with AI, and as a librarian, I can have a long conversation about those challenges, I also know that we cannot make AI disappear. Nor would I want to. We are in the uncomfortable time of major innovation upheaval, where we have to figure out how all of this fits within our lives and our societies. While we struggle now, we must also remember to be excited about the future we can create. Rather than refusing to use AI, we must seek to learn more about how it works and why people use it. I appreciate your evaluation of what it has been doing, but also the valid ways it may be tailored to help in the future.

Leave a Comment