AI Could Increase Research Output Faster Than Institutions Can Manage
Artificial intelligence is rapidly transforming scientific work, but experts warn it may unintentionally worsen existing academic pressures. Large language models allow researchers to generate papers, proposals, and reports at unprecedented speed.
Academics fear this acceleration may strain peer review systems, overwhelm funding agencies, and contribute to a growing imbalance between research creation and evaluation.
Incentive Systems Push Researchers Toward High-Volume Publishing Outputs
Current academic incentives reward researchers for publishing frequently, even if results are incremental. Experts say this pressure already fuels practices like splitting small findings into multiple papers.
AI tools make generating written content significantly easier, raising concerns that publication volume will increase even further and deepen existing structural challenges.
Funding Agencies Report Surges in AI-Written Grant Applications Worldwide
Research funders in Europe and beyond have reported dramatic increases in grant submissions that appear to be AI-assisted. Denmark’s science ministry recently described its agencies as being “run over” by applications.
Horizon Europe has also seen sharply lower success rates this year, with consultants pointing to a flood of large language model–generated submissions.
Recommended Article: AI May Deepen Global Inequality as Nations Struggle With Uneven…
Growing Paper Volume Raises Fears of Declining Novelty and Disruption
Analysts note that despite rising publication counts, research novelty may be declining. Studies show that papers and patents today connect less frequently across disparate scientific domains.
Experts argue that an AI-driven surge in incremental work could deepen this trend, making it harder to identify groundbreaking discoveries amid overwhelming volume.
Automated Screening Tools Risk Reinforcing Bias and Discouraging Innovation
Some organizations are testing AI to triage grant proposals or conduct preliminary reviews. One Spanish foundation used AI models to filter applications based on similarity to previously funded proposals.
Reviewers warned that such systems may disadvantage unconventional ideas, reduce diversity of thought, and unintentionally embed historical bias within key funding decisions.
Experts Caution Against Using AI to Solve Problems AI Created in Academia
A worst-case scenario, analysts say, is one where overwhelmed institutions rely heavily on AI to evaluate research, conduct meta-analyses, or screen papers. They warn this may further detach scientific judgment from human decision-making.
European Commission guidelines already advise funders to avoid using AI for proposal evaluations due to risks of hallucination, bias, and lack of transparency.

Reforms Needed to Ensure AI Supports Rather Than Overloads Research
Experts argue that meaningful system reforms are essential to prevent AI from accelerating unsustainable growth in research output. They recommend updating academic evaluation systems, redesigning peer review, and experimenting with innovative funding methods.
They emphasize that without structural change, AI could reinforce existing inefficiencies and push academic institutions toward a cycle of growing overload and diminishing research impact.












