Back to Basics: Why AI Disclosures in Publishing Should Center on Accountability, Not Reporting
This article is the second in a two-part series on AI disclosures in scholarly publishing. In Part 1, I argued that despite widespread AI use by researchers, author disclosure remains the exception rather than the norm. I explored why current disclosure guidelines are failing and why fear, ambiguity, and burden are driving AI use underground rather than making it transparent. In this follow-up, I turn to the more challenging question: what publishers should do about it.
The central mistake publishers make is assuming that more detailed reporting of AI use will lead to greater research integrity. In practice, the opposite seems to be happening. Overly demanding and vague disclosure requirements have created a transparency vacuum while doing little to protect what matters most in science: replication, reproducibility, and trust in the scholarly record.
The way forward is not more surveillance of authors’ workflows, but a renewed emphasis on author responsibility and outcome-based integrity.
The Wrong Question: “How Did You Use AI?”
Most current publisher policies implicitly ask authors to answer a deceptively simple question: How did you use AI?
This framing creates three immediate problems.
First, it assumes that AI use is a discrete, identifiable event rather than something increasingly embedded in everyday research tools. Second, it treats all AI use as equally relevant, when in fact AI assistance in writing prose presents fundamentally different risks than AI involvement in data analysis or modeling. Third, it shifts attention away from scholarly outcomes and toward process documentation: prompt logs, screenshots, and appendices that editors don’t have the time or expertise to evaluate.
In short, publishers are asking for information they cannot meaningfully use, while failing to ask for assurances they genuinely need.
The Right Question: “Can This Work Be Trusted and Reproduced?”
AI governance in publishing should be built around a much older and sturdier foundation: the norms of research integrity that long predate LLMs.
From that perspective, the relevant questions are not:
- Did the author use ChatGPT?
- What prompts did the author use?
- Do peer reviewers have familiarity and access to declared tools?
But rather:
- Are the data reliable?
- Are the methods transparent?
- Can the results be reproduced?
- Is authorship responsibility clearly assigned?
Reframing AI policy around these questions immediately clarifies where disclosure matters and where it does not.

A Practical Framework for Publishers
1. Replace Narrative Disclosures with Structured, Low-Burden Declarations
Rather than asking authors to “describe all AI use,” journals should implement simple, task-based declarations embedded directly into submission systems.
Authors should be asked to indicate categories of AI use (e.g., literature discovery, data analysis, code generation, language editing), not narrate workflows or share prompts. This standardization reduces ambiguity, minimizes burden, and creates consistent signals for editors without inviting overinterpretation. Crucially, such declarations should be routine and neutral, not framed as exceptional or suspicious.
2. Escalate Requirements Only Where Reproducibility is at Stake
Not all AI use poses the same risk to the scholarly record, and policies should state this explicitly.
AI use in data processing, statistical analysis, modeling, or code generation has direct implications for reproducibility and therefore warrants additional disclosure and justification. AI use in language editing, translation, or prose drafting generally does not.
Publishers should therefore adopt conditional disclosure according to editor discretion: deeper and more detailed disclosures should be required when AI touches analytic or methodological components. This keeps policy aligned with scholarly risk rather than distracting issues about writing assistance.
(PDF of table below available here)
| Stage of Research Workflow | Typical AI Use Cases | Author Declaration Required? | Editorial / Reviewer Scrutiny | Rationale |
| Idea Generation & Brainstorming | Generating research questions, refining hypotheses, exploratory prompts | No | None | Conceptual ideation does not affect reproducibility; responsibility lies in final design and execution. |
| Literature Discovery | AI-assisted search, summarization of papers, citation discovery tools | Optional / High-level | None | Equivalent in risk to advanced search tools; impact is indirect and non-determinative. |
| Study Design & Methods Planning | Drafting protocols, suggesting methodologies, experimental design support | Yes (high-level) | Moderate | Methods shape reproducibility; editors need assurance that final decisions were human-led. |
| Data Collection | AI-assisted instrumentation, automated data capture, image or signal preprocessing | Yes | Moderate to High | Data provenance and integrity must be transparent to assess downstream validity. |
| Data Cleaning & Preprocessing | Outlier detection, normalization, missing data handling | Yes | High | These steps materially affect results and must be reproducible. |
| Statistical Analysis & Modeling | AI-assisted statistical tests, model selection, parameter optimization | Yes (explicit) | High | Core analytical risk area; requires strong author responsibility and reproducibility assurances. |
| Code Generation or Debugging | Writing analysis scripts, simulation code, or pipelines | Yes (conditional) | High | Editors need confidence that code was reviewed, tested, and can be rerun independently. |
| Results Interpretation | AI-assisted explanation of findings, trend identification | Optional / Contextual | Moderate | Interpretation is scholarly judgment; concern arises only if AI substitutes for reasoning. |
| Figure & Image Generation | Creating plots, visualizations, or illustrative images | Yes (if data-derived) | Moderate | Visuals linked to data must be traceable and reproducible. |
| Writing, Editing & Translation | Drafting text, improving clarity, language editing, translation | No | None | Writing assistance does not affect data or methods; stylistic policing is inappropriate. |
3. Introduce an Explicit Author Responsibility Statement
At the heart of effective AI governance should be a simple but powerful declaration of accountability.
Authors should be required to affirm clearly and unambiguously that:
☐ I take full responsibility for all data, analyses, and interpretations in this manuscript, regardless of the tools used.
☐ I confirm that AI tools did not replace my judgment in analytical decisions.
☐ I attest that the work meets the journal’s standards for reproducibility and research integrity.
This approach mirrors existing practices around conflicts of interest and data availability. It avoids technological micromanagement while making responsibility unmistakable. Most importantly, it restores a principle that has become oddly obscured in AI debates: tools do not bear responsibility, authors do.
Educating Editors Matters as Much as Instructing Authors
One reason authors are reluctant to disclose AI use is not policy language, but editorial culture.
If authors believe that disclosure will quietly bias reviewers against them or signal lower rigor or competence, they will continue to hide AI use, regardless of how policies are written.
Publishers must therefore do something simple but significant: state clearly that AI use is not grounds for rejection.
Internally, editors and reviewers should be trained to evaluate AI-related concerns only insofar as they affect methodological soundness, data integrity, or reproducibility.
Editors and reviewers should be trained to ask:
- Does AI use affect methods validity?
- Does it obscure data provenance?
- Does it compromise replication?
This keeps AI evaluation where it belongs: inside scholarly assessment, not stylistic prejudice. Style, fluency, or perceived “AI-ness” of prose should be explicitly excluded from integrity judgments.
Carrots and Sticks: Incentivizing Responsible AI Use Without Policing
One objection to responsibility-based frameworks is that they “lack teeth.” This objection misunderstands both where enforcement should occur and how behavior actually changes in scholarly communities.
Sanctions should not be tied to using AI, but to failing core scholarly obligations. At the same time, publishers should recognize that transparency improves not only when poor behavior is punished, but when good behavior is visibly rewarded.
The “Carrot”: Normalizing and Recognizing Responsible Disclosure
Publishers have an opportunity to reframe AI transparency as a marker of good scholarly practice rather than a red flag. Even modest, symbolic recognition can shift author behavior meaningfully.
Examples of low-cost, high-signal incentives include:
- Standardized “Responsible AI Declaration” statements that appear alongside data availability or ethics statements, signaling good practice rather than suspicion.
- Editorial acknowledgments (visible to authors but not necessarily highlighted to reviewers) that thank authors for clear and complete AI-related declarations.
- Badging or labeling systems similar to “open data” or “transparent methods” indicators recognizing manuscripts that clearly articulate responsibility for AI-assisted analyses.
- Positive reinforcement in author guidelines, explicitly stating that transparent AI use is viewed as a strength, not a liability, when evaluating submissions.
None of these confer advantage in peer review, but all help normalize disclosure and reduce the perception that honesty carries risk.
The “Stick”: Proportionate, Outcome-Based Sanctions
Where sanctions are necessary, they should be proportionate and tied to scholarly outcomes and not to tool usage itself. A sensible enforcement regime might look like this:
- No sanction for undisclosed AI use in writing or language editing, absent evidence of misconduct or deception.
- Corrections or expressions of concern where AI involvement contributed to minor errors that do not invalidate results or conclusions.
- Retractions where AI use resulted in fabricated or hallucinated data, irreproducible analyses, falsified methods, or deceptive representation of results or provenance.
In such cases, AI is not the misconduct; it is merely the instrument. The violation remains what it has always been: a breach of research integrity. In cases where major research integrity issues are identified that include improper AI use, the matter should be pursued to the fullest extent.
Crucially, responsibility-based attestations give publishers a clear foundation for post-publication action without resorting to intrusive pre-submission surveillance, unreliable AI detection tools, or speculative judgments about authors’ workflows.
Why This Approach Encourages Transparency
When disclosure is framed as limited, routine, non-punitive, and clearly disconnected from stylistic judgment, authors have far less incentive to hide AI use. Transparency increases not because authors are forced to comply, but because honesty no longer feels risky.
This mirrors what we have learned slowly and imperfectly from data sharing, preregistration, and open methods: compliance follows clarity, incentives, and trust, not ever-expanding documentation requirements.
Conclusion: Governing Outcomes, Not Tools
AI has exposed weaknesses in scholarly publishing but it did not create them. The current disclosure impasse is less about technology than about misaligned incentives and misplaced attention.
Publishers do not need to legislate every interaction between authors and machines. They need to reaffirm something simpler and more durable: that authors are accountable for the work they submit, that reproducibility remains non-negotiable, and that transparency should be rewarded rather than punished.
The path forward lies not in tracking AI use more aggressively, but in doubling down on the foundations of scholarly trust that AI has not replaced and cannot.
Author’s note: I wrote a rough draft, including a mix of bullet points and more fleshed-out narrative paragraphs. I then ask GPT to turn my rough ideas into an essay draft form. I then reviewed the output, made substantive revisions, sent to colleagues for feedback, and then sent it off for publication.
Discussion
15 Thoughts on "Part 2 — Why Authors Aren’t Disclosing AI Use and What Publishers Should (Not) do About It"
I really appreciate this emphasis on accountability and outcomes rather than exhaustive reporting of AI workflows. One challenge I see editors struggling with is that policies are often written as if all journals have the same editorial capacity, reviewer expertise, and governance structures—which simply isn’t the case.
I’ve been exploring this issue through what I call the Journal Systems Framework, which looks at journals as systems operating under different resource and oversight conditions. From that perspective, disclosure models that rely on author responsibility and reproducibility are far more robust than ones that assume editors can meaningfully audit prompts, tools, or workflows.
This piece does a great job of articulating why more reporting is not the same as more integrity.
(https://writers-camp.org/2026/01/15/how-journals-work-systems-based-framework-scholarly-publishing/
Thanks Leslie. Out of curiosity, do you think publishers should create standard protocols and guidelines across their journal portfolios or do you think each editor should decide policy for their specific journal?
That’s a great question—and my answer is “both, but with clear boundaries.”
At the publisher level, I do think there’s value in setting shared principles and guardrails: expectations around accountability, authorship responsibility, ethical use, and what editors should be able to defend if a decision is questioned later. Those baseline norms help with consistency and credibility across a portfolio.
Where I get concerned is when those principles turn into highly prescriptive, workflow-heavy protocols that assume all journals have the same editorial capacity, reviewer expertise, and governance infrastructure. In practice, that’s simply not true—even within a single publisher’s portfolio. When policies are written as if every journal operates like a flagship title, they can become unworkable or even counterproductive for smaller, capacity-constrained journals.
So my view is that publishers should articulate what needs to be upheld (accountability, transparency proportional to risk, editorial responsibility), while editors need discretion over how those principles are implemented in their specific context. That flexibility is essential if we want policies that are actually followed—and that support good editorial judgment rather than replacing it with box-checking.
In the end, AI policies should strengthen editorial decision-making, not create another layer of performative compliance that varies wildly in feasibility from one journal to the next.
Great article! The table appears to cut off on the right? Is there a way to post full table as contains helpful information.
Really great article and gives me a lot to think about. Curious to hear your thoughts on hallucinated references and what these indicate might have happened – for me this suggests the writer hasn’t engaged with the material and raises more doubts about the legitimacy of the rest of the work. I’d love to hear others thoughts on this though!
Thanks Kim, I appreciate it!
I would say that hallucinated references are a ‘Signal’ similar to many other telltale signs of AI, such as tortured phrases or generic, bland writing. However, there is the possibility that the author made an honest mistake and the article is still worthy of consideration. For example, I once tried to use AI to help sort out formatting for an article and it slipped in a hallucinated reference very subtly. Thankfully, I caught it before sending out for publication. I’m happy to discuss this further if you’d like…
Such a great read! Context is so important when declaring AI use, especially when it is so deeply embedded at every step of the research process. Also agree about the importance of educating editors and reviewers about AI disclosures
Avi, If researchers are required to vouch for the integrity of their work, shouldn’t we apply that standard to the discussion of the standard itself?
That said, I ran this article through Pangram Labs’ AI detection tool. Result: 100% AI-generated. (I’d upload the screenshot, but there’s no upload feature on this site.)
Pangram’s detection isn’t an accusation. Rather, it’s meant to serve as an observation that proves your own thesis. An article about AI disclosure, read by people who write disclosure policy, published without AI disclosure, reveals a transparency gap that’s now fully visible.
Thanks for your comment Jayne.
I’d be happy to share the workflow I use for transparency and for others to learn from for some of the different writing tasks I complete. I write a rough draft, which is usually a mix of bullet points and more fleshed-out narrative paragraphs. I then ask GPT to turn my rough ideas into essay draft form. I then review the output, make substantive revisions, send to colleagues for feedback and then send off for publication.
So Pangram’s assessment is somewhat correct and somewhat incorrect. It is true that many of the words are AI-generated, but there is a significant amount that wasn’t as well. And regardless, I sign off on every word so I actually don’t think it makes much difference.
I’m happy to add a disclaimer, but if you read both of my articles carefully, you will know that my argument is that disclosure should only be relevant is it substantially affects the reliability or ability to reproduce the results. Being that this is an op-ed editorial, I don’t think either of those applies.
I encourage you to see the continuation of my argument on Linkedin here: https://www.linkedin.com/posts/avistaiman_what-if-i-told-you-that-most-researchers-activity-7430177250114916352-8GM3?utm_source=share&utm_medium=member_desktop&rcm=ACoAAAh6xQ8BfFCvKvx3mr6a8WsN1AggdMT3uL4
That would be grounds for rejection at our journal. We state no AI assisted writing. We are a mid ranked scholar-led critical social sciences journal with no APC and a huge number of submissions, and grappling with workload. There is little STEM content where ‘data’ can be analysed by a computer, unlike many more scientific journals. We do not want articles with anything more than assisted grammar checking, after 30 years of just having authors write their own work. Authors will be directed elsewhere.
Why?
Thanks for sharing your AI work process. Mine’s similar: I write the narrative, iterate with AI, meaning I use AI to research, analyze information, brainstorm, understand arguments and refine, then verify facts independently. Depending on the piece, I’ll also ask a subject matter expert to verify whether what I’ve written is accurate. I realize that AI detection tools aren’t perfect, but I’d rank Pangram as one of the best. That said, some detectors, like Pangram, flag AI-polished language. Why is that different from human copy editing? That’s what human copy editors do, and I was a copy editor for The Washington Post. I always honored a reporters words. You can ask AI to honor your words as well. I wish we could change the discourse to hybrid intelligence.
Thank you so much for this scholarly kitchen newsletter. Would you please give me concise note about pros and cons of Pangram AI detector tool for daily utilization of academics and also what makes different from others AI detectors.
From my experience, Pangram has been more accurate. My nonfiction book, Act Early Against Autism, was published in 2008 by Penguin/Perigee. I took 10,000 words (the maximum allowed at the time) and ran them through Originality.ai. It came back as 65% AI-detected, a major false positive. I should note that this was over a year ago, and I’d like to think that Originality.ai has since improved.
Act Early is not part of the Anthropic lawsuit about using pirated book copies to train LLMs. When I ran the same sample through Pangram, it came back as Human Authored. I haven’t revisited Originality.ai since. I’ve also used GPTZero, though not with Act Early. In that case, I generated around 100 words with AI and then ran them through its Humanizer. Even with a Humanizer, you still need to manually edit the copy because it tends to repeat phrases or an entire sentence. After cutting duplicate language and cleaning up text it awkwardly reworded, I re-ran the text through both Pangram and GPTZero. Each scored it as Human Authored, though GPTZero flagged about 6% as AI-generated.
Overall, I’d say the piece was still largely AI-generated, but my edits clearly disrupted the statistical patterns that marked it as such, even if I can’t claim to have written it word for word.