AI boosts scientific productivity but erodes paper quality

A Cornell University study reveals that AI tools like ChatGPT have increased researchers' paper output by up to 50%, particularly benefiting non-native English speakers. However, this surge in polished manuscripts is complicating peer review and funding decisions, as many lack substantial scientific value. The findings highlight a shift in global research dynamics and call for updated policies on AI use in academia.

Since ChatGPT's widespread adoption in late 2022, scientists have reported higher productivity, with journal editors noting an influx of well-written but low-value submissions. A Cornell study, published on December 18, 2025, in Science, analyzed over 2 million preprints from arXiv, bioRxiv, and SSRN, spanning January 2018 to June 2024. Researchers developed a detector to identify LLM-assisted papers by comparing them to pre-2023 human-written ones.

The results show a clear productivity boost: authors likely using LLMs posted about one-third more papers on arXiv and over 50% more on bioRxiv and SSRN. The gains were most pronounced for non-native English speakers, with researchers from Asian institutions increasing output by 43% to 89.3%, depending on the platform. "It is a very widespread pattern, across different fields of science," said Yian Yin, assistant professor of information science at Cornell's Ann S. Bowers College of Computing and Information Science.

Beyond writing, AI search tools like Bing Chat improved literature reviews by surfacing newer, more diverse sources. First author Keigo Kusumegi noted, "People using LLMs are connecting to more diverse knowledge, which might be driving more creative ideas."

Yet, challenges emerge in evaluation. Human-written papers with complex language often signal quality and higher journal acceptance rates. In contrast, LLM-assisted papers, despite sophisticated prose, are less likely to be accepted, suggesting that polish no longer reliably indicates value. This disconnect could hinder editors, reviewers, and funders, as raw publication counts become misleading.

The observational study calls for experimental follow-ups and policy updates. Yin is hosting a symposium on March 3-5, 2026, in Ithaca to discuss AI's role in research. Co-authors include Xinyu Yang, Paul Ginsparg, Mathijs de Vaan, and Toby Stuart; funding came from the National Science Foundation.

As AI evolves into a "co-scientist," Yin emphasizes transparency: "The question is, how exactly have you used AI and whether it's helpful or not."

Relaterede artikler

Illustration of Swedes in a Stockholm cafe using AI chatbots amid survey stats on rising usage and skepticism.
Billede genereret af AI

Øget brug af AI-chatbots blandt svenskere – men også bekymringer

Rapporteret af AI Billede genereret af AI

Ifølge den seneste SOM-undersøgelse fra Göteborgs universitet er andelen af svenskere, der ugentligt chatter med en AI-bot, steget fra 12 til 36 procent mellem 2024 og 2025. Samtidig er skepsissen over for AI vokset, hvor 62 procent ser det som en større risiko end mulighed for samfundet.

At the American Physical Society Global Physics Summit in Denver, Colorado, thousands of researchers are using AI chatbots to simplify complex talks. The event has sparked intense discussions on whether artificial intelligence will transform physics research. Speakers presented contrasting views on AI's potential and limitations.

Rapporteret af AI

A new study from Brown University identifies significant ethical concerns with using AI chatbots like ChatGPT for mental health advice. Researchers found that these systems often violate professional standards even when prompted to act as therapists. The work calls for better safeguards before deploying such tools in sensitive areas.

A study applying Chile's university entrance exam, PAES 2026, to AI models shows several systems scoring high enough for selective programs like Medicine and Civil Engineering. Google's Gemini led with averages near 950 points, outperforming rivals like ChatGPT. The experiment underscores AI progress and raises questions about standardized testing efficacy.

Rapporteret af AI

A new research paper argues that AI agents are mathematically destined to fail, challenging the hype from big tech companies. While the industry remains optimistic, the study suggests full automation by generative AI may never happen. Published in early 2026, it casts doubt on promises for transformative AI in daily life.

A CNET commentary argues that describing AI as having human-like qualities such as souls or confessions misleads the public and erodes trust in the technology. It highlights how companies like OpenAI and Anthropic use such language, which obscures real issues like bias and safety. The piece calls for more precise terminology to foster accurate understanding.

Rapporteret af AI

Superhuman, the company behind writing tool Grammarly, has disabled its Expert Review feature following complaints and a class action lawsuit. The tool used AI to generate writing feedback attributed to famous authors and academics without their consent. CEO Shishir Mehrotra announced the shutdown on March 11, 2026.

søndag d. 22. marts 2026, 03.30

Spanish Congress deputies use AI to prepare speeches

mandag d. 2. marts 2026, 04.22

Japan shows high AI trust despite low workplace use

torsdag d. 26. februar 2026, 23.44

Study shows AI can deanonymize online users from posts

lørdag d. 21. februar 2026, 01.40

Generative AI outperforms human teams in analyzing medical data

mandag d. 16. februar 2026, 03.16

Folha columnists debate AI use in writing

tirsdag d. 3. februar 2026, 09.31

Senior OpenAI staff leave amid ChatGPT focus

søndag d. 18. januar 2026, 01.24

AI companies gear up for ads as manipulation threats emerge

torsdag d. 15. januar 2026, 10.16

AI models risk promoting dangerous lab experiments

onsdag d. 24. december 2025, 04.08

How AI coding agents function and their limitations

søndag d. 14. december 2025, 21.07

OpenAI report shows AI saves workers under an hour daily

 

 

 

Dette websted bruger cookies

Vi bruger cookies til analyse for at forbedre vores side. Læs vores privatlivspolitik for mere information.
Afvis