Author: Denis Avetisyan
A new analysis reveals a growing presence of AI-authored text within open-access medical literature, raising questions about originality and transparency.

Longitudinal study shows increasing detection of AI-generated content, particularly in commentaries and global health publications, with a significant disparity between use and author disclosure.
Despite increasing concerns regarding academic integrity and the validity of published research, the extent to which generative artificial intelligence is being utilized in medical literature remains largely unknown. This study, ‘Rising Prevalence of Detected AI-Generated Text in Medical Literature: Longitudinal Analysis in Open Access Articles’, analyzed over 7,000 articles published in JAMA Network Open from 2022-2025 and found a significant increase in detectable AI-generated content, rising from 0.0% to 11.3% over the study period. Notably, Invited Commentaries and publications focused on Global Health demonstrated the highest proportions of AI-generated text, with a substantial gap between detected use and author disclosures. As these tools become more sophisticated, how will we ensure the reliability and transparency of medical knowledge?
The Erosion of Scholarly Provenance: A Necessary Reckoning
The integration of Large Language Models into academic writing practices is prompting significant discussions regarding the fundamental principles of originality and transparency within research. These powerful AI tools, capable of generating human-quality text, present challenges to traditional methods of assessing authorship and intellectual contribution. While offering potential benefits in streamlining the writing process, their use raises concerns about the potential for plagiarism, the blurring of lines between human and machine-generated content, and the overall integrity of published research. This shift necessitates a re-evaluation of how academic communities define and verify originality, demanding new strategies to ensure accountability and maintain public trust in scientific findings. The increasing sophistication of these models further complicates the matter, requiring continuous development of methods to distinguish between authentically authored work and AI-assisted composition.
The escalating integration of artificial intelligence into academic writing is challenging the foundations of scholarly publishing, as many journals currently lack the sophisticated tools necessary to consistently differentiate between human-authored and AI-generated content. This deficiency isn’t merely a technical hurdle; it represents a growing crisis of trust within the scientific community. Without reliable detection methods, the potential for plagiarism, compromised research integrity, and the dissemination of inaccurate or fabricated findings increases substantially. The absence of standardized protocols for identifying AI’s influence undermines peer review processes and casts doubt on the validity of published research, potentially eroding public confidence in scientific advancements and the institutions that produce them. Addressing this vulnerability is paramount to preserving the credibility and reliability of scholarly work in an era of increasingly accessible and powerful language models.
A comprehensive analysis of 7251 scholarly articles uncovered a notable presence of artificially generated text, with 2.7% demonstrating clear indicators of AI authorship. This finding underscores a critical gap in current academic publishing practices, as existing methods often fail to reliably identify content produced by large language models. The detection of AI-generated text is no longer a hypothetical concern but an immediate necessity for maintaining the integrity and trustworthiness of scientific literature, prompting a growing demand for effective and widely accessible detection tools to safeguard originality and ensure accountability within the research landscape.
A striking surge in the utilization of artificial intelligence within academic publishing is becoming increasingly apparent. Analysis reveals that manuscripts incorporating AI-generated text have risen dramatically, moving from a negligible 0.0% in January 2022 to 11.3% by March 2025. This exponential growth underscores a fundamental shift in scholarly writing practices, suggesting that AI is no longer a peripheral tool but an increasingly integrated component of manuscript creation. The rapid acceleration of AI influence demands a critical evaluation of current publishing standards and the implementation of effective detection mechanisms to maintain the integrity and trustworthiness of scientific literature.

Methodological Rigor: Data Acquisition and Analysis
Data acquisition was performed using web scraping techniques implemented in Python. The Beautiful Soup library was utilized to parse HTML content retrieved from JAMA Network Open. This systematic approach allowed for the automated collection of manuscript data, including both acknowledgment sections and full text, circumventing manual data entry and enabling a large-scale analysis. The scraping process was designed to adhere to the website’s terms of service and robots.txt file, respecting crawl delays to minimize server load. Collected data was stored in a structured format suitable for subsequent analysis and processing.
Data extraction prioritized two distinct manuscript components: the acknowledgment section and the full text. Acknowledgment sections were targeted due to the higher probability of authors voluntarily disclosing the use of artificial intelligence tools in their research process. The full text of each manuscript was also extracted to enable a comprehensive analysis using an AI detection tool, allowing for the assessment of potential AI-generated content beyond what was explicitly stated in the acknowledgments. This dual-focus approach facilitated both the identification of self-reported AI use and an independent evaluation of the manuscripts’ content for AI-assisted writing.
Regular expression matching served as the initial method for identifying disclosures of AI use within the acknowledgment sections of manuscripts. This technique involved constructing pattern-based searches to locate explicit statements indicating the application of artificial intelligence or machine learning tools during the research or writing process. Identified statements were then manually reviewed to confirm relevance and accuracy. The resulting dataset of acknowledged AI use cases established a quantifiable baseline against which the findings from the commercial AI detection tool, Originality.AI, could be compared, allowing for an assessment of its sensitivity and specificity in identifying AI-generated content.
Originality.AI, a commercially available text analysis tool, was utilized to assess the full text of each manuscript for content potentially generated by large language models. This tool employs a proprietary algorithm to identify patterns and characteristics indicative of AI authorship, providing a percentage-based score reflecting the likelihood of AI involvement. Manuscripts were processed individually through the platform, and scores were recorded for quantitative analysis. The tool’s detection methodology focuses on identifying statistical anomalies in perplexity, burstiness, and stylistic consistency, features that often differentiate human-written text from AI-generated content. It is important to note that Originality.AI, like all AI detection tools, is not infallible and provides a probability assessment, not a definitive determination, of AI authorship.
Empirical Evidence: Quantifying AI’s Presence in Scholarly Work
Analysis of manuscripts published in JAMA Network Open demonstrated the presence of AI-generated text across various publication types. While the proportion of AI-generated content varied, detectable instances were confirmed within the analyzed dataset. Specifically, 6.7% of Invited Commentaries, 2.2% of Original Investigations, and 1.4% of Research Letters contained AI-generated text. Further analysis revealed that Global Health articles exhibited the highest prevalence of AI-generated content, reaching 8.6%. However, disclosure of AI use remained low, with only 0.2% of all analyzed articles reporting the utilization of large language models.
Chi-Squared testing revealed statistically significant variations in the detection rates of AI-generated text based on publication type and subject domain within the JAMA Network Open corpus. Specifically, Invited Commentaries exhibited a substantially higher proportion of AI-generated content (6.7%) compared to Original Investigations (2.2%) and Research Letters (1.4%). Analysis by subject area indicated that Global Health articles had the highest prevalence of AI-generated content (8.6%). These differences suggest that the utilization of AI writing tools is not uniform across all content types and research areas published within the journal, potentially reflecting varying needs or practices among different author groups.
Statistical analysis using Mann-Kendall testing demonstrated a significant increasing trend in AI detection rates over the study period (p << 0.001). This indicates a growing presence of AI-generated text within the analyzed manuscripts over time. The statistically significant positive trend suggests an increasing adoption rate of AI writing tools, potentially reflecting their growing accessibility and integration into academic writing workflows. The consistently low p-value further supports the robustness of this observed trend, minimizing the likelihood of it occurring due to random chance.
Analysis of manuscripts published in JAMA Network Open revealed varying rates of AI-generated text based on publication type. Specifically, 6.7% of Invited Commentaries were identified as containing AI-generated content, representing the highest proportion among the examined categories. Original Investigations demonstrated a lower prevalence, with 2.2% containing AI-generated text. Research Letters exhibited the lowest rate, at 1.4%. These findings suggest that AI writing tools are utilized at different frequencies depending on the specific genre and purpose of the published work.
Analysis of manuscripts published in JAMA Network Open revealed significant disparities in the presence of AI-generated text across different article types. Specifically, articles categorized as Global Health exhibited the highest proportion of AI-generated content, reaching 8.6% of all manuscripts in that category. This contrasts sharply with the overall disclosure rate for AI usage across all analyzed articles, which was only 0.2%. This indicates a substantial gap between the actual utilization of AI writing tools, particularly within Global Health publications, and the reporting of that usage to readers and the scientific community.
Analysis of disclosed Large Language Model (LLM) use revealed a discrepancy between author reporting and AI detection results. Of the 15 articles where LLM use was explicitly stated, 6 (40.0%) were subsequently classified as containing greater than 10% AI-generated text based on detection tools. Conversely, only 6 out of the 195 articles flagged as containing AI-generated content (3.1%) included author disclosure of LLM use, indicating a significant gap between actual AI utilization and transparent reporting within the analyzed dataset.
A Crisis of Integrity: Implications and Future Directions
Recent increases in the sophistication and accessibility of artificial intelligence tools necessitate a re-evaluation of academic integrity standards. The evolving landscape demands that institutions and publishers move beyond simply prohibiting AI-generated content and instead develop nuanced policies that address permissible uses, appropriate disclosure, and the ethical boundaries of collaboration with these technologies. Clear guidelines are crucial not only to deter academic misconduct, but also to equip students and researchers with a framework for responsible innovation and to foster a culture of transparency in scholarly work. Without such clarity, the potential benefits of AI in research – such as accelerating literature reviews or assisting with data analysis – risk being overshadowed by concerns regarding authorship, originality, and the validity of published findings. Establishing these policies proactively will be essential for maintaining the trust and credibility of academic institutions and the research they produce.
Academic journals and educational institutions now face a critical juncture, requiring proactive strategies to navigate the complexities introduced by AI-generated content. The increasing sophistication of large language models necessitates a re-evaluation of existing policies concerning authorship, originality, and plagiarism. Simply banning AI tools proves insufficient; instead, a nuanced approach focusing on responsible AI integration is crucial. This includes developing clear guidelines for AI-assisted writing, establishing robust methods for verifying the provenance of submitted work, and fostering a culture of academic integrity that values transparency and ethical scholarship. Furthermore, institutions must invest in training for both faculty and students regarding the appropriate use of these technologies, ensuring a shared understanding of the ethical boundaries and methodological considerations involved. Failure to adapt risks undermining the credibility of scholarly publishing and the validity of academic credentials.
Current AI detection tools, while increasingly prevalent, remain imperfect and necessitate continued refinement. A significant challenge lies in minimizing false positives – the incorrect identification of human-written text as AI-generated – which can have serious consequences for students and researchers. Ongoing research focuses on developing more nuanced algorithms that consider stylistic features, contextual understanding, and the subtle markers of human authorship, moving beyond simple pattern matching. This includes exploring techniques like watermarking, which subtly embeds identifiable signals within AI-generated text, and adversarial training, where detection tools are challenged with increasingly sophisticated AI outputs to improve their robustness. Ultimately, enhancing the precision of these tools is crucial not only for upholding academic integrity but also for fostering a fair and trustworthy research environment.
The advent of Large Language Models (LLMs) like GPT-4o necessitates a thorough re-evaluation of established academic practices. These models, capable of generating coherent and contextually relevant text, present both opportunities and challenges for research. While LLMs can assist with literature reviews, data analysis, and even hypothesis generation, their use demands careful consideration of authorship, originality, and potential biases embedded within the algorithms. A robust, open discussion is crucial to establish guidelines for appropriate LLM integration-determining when such tools constitute legitimate research assistance versus academic misconduct. Furthermore, the evolving capabilities of these models require ongoing assessment of existing detection methods and the development of new strategies to uphold the integrity and validity of scholarly work. Ignoring the potential of – and pitfalls associated with – LLMs risks both stifling innovation and compromising the foundations of evidence-based knowledge creation.
The study’s findings regarding the escalating presence of AI-generated text in medical publications necessitate a rigorous re-evaluation of originality standards. The observed discrepancy between detected AI usage and authorial disclosure introduces a critical flaw in the verification process-a flaw demonstrable through asymptotic analysis of reporting accuracy. As Ada Lovelace observed, “The Analytical Engine has no pretensions whatever to originate anything.” This holds true even now; these large language models merely rearrange existing information. The core issue isn’t the tool’s existence, but the failure to acknowledge its influence, undermining the provable correctness expected within medical literature. The increasing prevalence demands a shift towards verifiable provenance, not simply reliance on authorial statements.
What’s Next?
The observed increase in algorithmically generated text within medical literature isn’t merely a quantitative curiosity; it’s a challenge to the very notion of scholarly contribution. The study highlights a disturbing asymmetry: detection capabilities demonstrably outpace authorial transparency. This suggests that, presently, reliance on honesty is a flawed assumption. Future work must move beyond simply identifying AI’s presence and concentrate on developing metrics for assessing the impact of such text on scientific validity. A reproduction crisis looms if conclusions are built on foundations that lack verifiable provenance.
A critical, and largely unaddressed, issue is the deterministic nature of truth. If a result cannot be reproduced-or, more precisely, if the process leading to a result is obscured by opaque algorithmic contribution-its reliability is fundamentally compromised. The current emphasis on detection is, in a sense, treating a symptom, not the disease. The focus should shift to creating systems where algorithmic contribution is either explicitly acknowledged and auditable, or demonstrably absent.
The divergence observed in publication types – particularly within Invited Commentaries and Global Health – warrants deeper investigation. Are these areas more susceptible to this practice due to pressures for rapid publication, or do they represent a qualitatively different use case? Ultimately, the question isn’t whether algorithms can contribute to medical discourse, but whether they do so in a manner consistent with the principles of rigorous, verifiable, and therefore, meaningful, scientific inquiry.
Original article: https://arxiv.org/pdf/2603.19316.pdf
Contact the author: https://www.linkedin.com/in/avetisyan/
See also:
- Physics Proved by AI: A New Era for Automated Reasoning
- Invincible Season 4 Episode 4 Release Date, Time, Where to Watch
- American Idol vet Caleb Flynn in solitary confinement after being charged for allegedly murdering wife
- Gold Rate Forecast
- “Wild, brilliant, emotional”: 10 best dynasty drama series to watch on BBC, ITV, Netflix and more
- eFootball 2026 is bringing the v5.3.1 update: What to expect and what’s coming
- Total Football free codes and how to redeem them (March 2026)
- Seeing in the Dark: Event Cameras Guide Robots Through Low-Light Spaces
- Magicmon: World redeem codes and how to use them (March 2026)
- Hatch Dragons Beginners Guide and Tips
2026-03-23 17:31