Researchers at Cornell University have found that the use of large language models (LLMs) like ChatGPT is significantly boosting productivity in scientific research, particularly for non-native English speakers. Their study highlights a growing trend since the release of ChatGPT in late 2022, which has led to an increase in the number of manuscripts submitted to various preprint platforms. However, this uptick in productivity is raising concerns about the quality and originality of scientific work, as many journal editors report a rise in submissions that appear well-written but lack substantial scientific value.
The findings, published in a paper titled “Scientific Production in the Era of Large Language Models” in *Science* on December 18, reveal a notable shift in the ecosystem of scientific publishing. Yian Yin, assistant professor of information science at Cornell, pointed out that this trend spans multiple disciplines, including physical, computer, biological, and social sciences. “There’s a big shift in our current ecosystem that warrants a very serious look, especially for those who make decisions about what science we should support and fund,” he stated.
To gauge the impact of LLMs on scientific output, Yin’s team examined over 2 million papers posted from January 2018 to June 2024 across three prominent preprint platforms: arXiv, bioRxiv, and the Social Science Research Network (SSRN). By comparing papers believed to be written by humans with those likely generated using LLM assistance, the researchers developed a model to flag AI-influenced texts. This model enabled them to estimate how many papers authors submitted before and after adopting LLMs and whether these papers were subsequently accepted by journals.
The results indicated a clear productivity boost linked to LLM usage. On arXiv, researchers identified as using LLMs submitted approximately one-third more papers than their peers who did not. On bioRxiv and SSRN, the increase was even more pronounced, exceeding 50%. Notably, the most significant gains were observed among scientists who write in English as a second language. Researchers affiliated with Asian institutions, for instance, increased their output by between 43% and 89% after adopting LLMs, depending on the platform. Yin anticipates this technological advantage may eventually alter global scientific productivity patterns, particularly benefitting regions previously hampered by language barriers.
The study also uncovered potential advantages concerning literature searches and citation practices. Bing Chat, regarded as the first widely used AI-powered search tool, outperformed traditional search engines in identifying newer and relevant papers. Traditional tools often favored older, more frequently cited sources. “People using LLMs are connecting to more diverse knowledge, which might be driving more creative ideas,” noted first author Keigo Kusumegi, a doctoral student in information science, who plans to explore the relationship between AI usage and the emergence of innovative, interdisciplinary science.
Despite the productivity benefits, the findings raise critical concerns regarding the peer review process. Historically, complex writing styles characterized by longer sentences and advanced vocabulary have been indicators of high-quality research. However, the study revealed that papers likely written with LLM assistance did not follow this trend. Even when these AI-generated papers excelled in writing complexity, they faced lower acceptance rates from journals. Yin interprets this as an indication that the quality of writing alone may no longer be a reliable measure of scientific merit, complicating the decisions of editors and reviewers who are tasked with identifying valuable contributions to the field.
This disconnect could have far-reaching implications, suggesting that institutions and funding agencies may need to reassess their reliance on raw publication counts as a gauge of scientific significance. Moving forward, the researchers aim to delve deeper into the cause-and-effect relationship of LLM usage in scientific writing through controlled experiments. Yin is also organizing a symposium set for March 3-5, 2026, at Cornell’s Ithaca campus, focusing on the implications of generative AI in research and what policymakers should consider as these tools become increasingly integrated into the scientific process.
Yin concludes that as AI tools become more ubiquitous in writing, coding, and idea generation, their influence will only grow. The pressing question for researchers and institutions is shifting from whether AI is used to how it is being utilized effectively. “Already now, the question is not, have you used AI? The question is, how exactly have you used AI and whether it’s helpful or not,” he stated.
See also
HCLTech Joins Microsoft Discovery to Boost Agentic AI in Scientific Research
AI Race Shifts: Companies Prioritize Adoption Over Model Quality as Spending Scrutinized
DOE Launches ‘Genesis Mission’ to Double U.S. Research Productivity with AI by 2035
AI Researcher R. Xie Reveals Deep Learning’s Role in BRI Green Production Optimization
AI Models in Economic Game Show Key Differences from Human Strategic Thinking


















































