Logo
Daily Brief
Following
Why
The AI Science Rush

The AI Science Rush

Large language models earn recognition for frontier research—and expose cracks in scientific publishing

Overview

Science magazine named large language models doing frontier science a runner-up breakthrough of 2025. DeepMind's Gemini earned a gold medal at the International Mathematical Olympiad—a feat forecasters in 2021 predicted wouldn't happen until 2043. OpenAI's GPT-5 optimized a gene-editing protocol and achieved a 79x efficiency gain in a real wet lab. Google's AI co-scientist reproduced years of human research in days.

The recognition caps a gold rush that's channeling hundreds of millions into AI science startups like Periodic Labs ($300M seed) and Lila Sciences ($1.3B valuation). But it's also exposing a crisis: Scientists using LLMs now publish 50% more papers, yet paper mill fraud and AI-generated junk are overwhelming peer review. The question isn't whether AI can do science—it's whether the system can handle what comes next.

Key Indicators

79x
Gene editing efficiency gain (GPT-5)
OpenAI's model optimized a molecular cloning protocol to recover 79 times more sequence-verified clones than baseline
50%+
Research output increase for LLM users
Scientists flagged as using LLMs posted 33-50% more papers on arXiv, bioRxiv, and SSRN
$1.3B
Lila Sciences valuation
AI science startup raised $550M total to build 'scientific superintelligence' with automated labs
10,000+
Paper retractions (2023)
Retractions quadrupled in 20 years, majority due to misconduct; 1 in 50 papers show paper mill patterns
53%
Researchers using AI in peer review
More than half of academics now use AI tools when evaluating manuscripts, often against guidance

People Involved

Demis Hassabis
Demis Hassabis
CEO and Co-founder, Google DeepMind (Co-winner of 2024 Nobel Prize in Chemistry)
Ekin Dogus Cubuk
Ekin Dogus Cubuk
Co-founder, Periodic Labs (Leading $300M AI science startup)
Gary Peltz
Gary Peltz
Liver Disease Researcher, Stanford Medicine (Testing Google's AI co-scientist for drug repurposing)

Organizations Involved

Google DeepMind
Google DeepMind
AI Research Division
Status: Leading AI scientific discovery breakthroughs

Alphabet's AI research lab pioneering large language models for scientific discovery across mathematics, biology, and materials science.

PE
Periodic Labs
AI Science Startup
Status: Building automated scientific discovery platform

AI startup automating scientific discovery, initially targeting superconductor invention.

LI
Lila Sciences
AI Science Startup
Status: Building scientific superintelligence with automated labs

Startup pairing specialized AI models with automated laboratories to accelerate biological research.

SA
Sakana AI
AI Research Startup
Status: Developed first peer-reviewed AI-generated research paper

Tokyo-based startup that created The AI Scientist system for fully automated research.

SC
Science Magazine
Scientific Journal
Status: Recognized LLMs as 2025 Breakthrough runner-up

Peer-reviewed academic journal published by the American Association for the Advancement of Science.

Timeline

  1. Study Shows 50% Output Boost, Quality Concerns

    Research Publication

    Cornell study finds LLM users publish 33-50% more papers, but AI-polished work less likely to be accepted.

  2. Science Names LLMs Breakthrough Runner-Up

    Recognition

    Science magazine recognizes large language models doing frontier science as 2025 Breakthrough of the Year runner-up.

  3. GPT-5 Achieves 79x Lab Efficiency Gain

    Research Breakthrough

    OpenAI announces GPT-5 optimized gene-editing protocol in real wet lab, introducing novel enzyme mechanism.

  4. Lila Sciences Hits $1.3B Valuation

    Funding

    AI lab raises $115M extension from Nvidia, General Catalyst, bringing total raised to $550M for automated laboratories.

  5. Periodic Labs Raises $300M Seed

    Funding

    Former OpenAI and DeepMind researchers raise one of largest seed rounds ever from a16z, Nvidia, Bezos.

  6. AI Co-Scientist Validates Drug Candidates

    Research Validation

    Stanford publishes findings: Google's AI co-scientist identified liver fibrosis drugs in days versus years of human research.

  7. Gemini Wins Gold at Math Olympiad

    Competition Victory

    DeepMind's Gemini Deep Think scores 35/42 points, achieving gold medal standard at International Mathematical Olympiad.

  8. First AI-Generated Peer-Reviewed Paper

    Publication

    Sakana's AI Scientist v2 produces first entirely AI-generated paper accepted at ICLR workshop after peer review.

  9. First AI-Enabled Nobel Prize

    Recognition

    Hassabis and Jumper win Nobel Prize in Chemistry for AlphaFold—first Nobel recognizing AI-enabled scientific discovery.

  10. Sakana Releases The AI Scientist

    Technology Release

    Sakana AI unveils open-source system automating entire research lifecycle from hypothesis to paper writing.

  11. AlphaFold Database Launched

    Platform Release

    DeepMind releases open-source AlphaFold and database with 360,000 protein structures; now exceeds 200 million predictions.

  12. AlphaFold2 Solves Protein Folding

    Scientific Breakthrough

    DeepMind's AlphaFold2 predicts protein structures with unprecedented accuracy, solving a 50-year grand challenge in biology.

Scenarios

1

AI Scientists Become Standard Research Partners by 2027

Discussed by: Google Research, Stanford Medicine researchers, pharmaceutical industry analysts

LLMs evolve from hypothesis generators to full research collaborators integrated into every lab. Isomorphic Labs and similar companies advance multiple AI-designed drug candidates through clinical trials. Research productivity accelerates dramatically, particularly for non-English-native scientists. The shift resembles how computational modeling became standard in the 1990s—initially controversial, ultimately indispensable. Traditional wet-lab work continues but AI handles literature review, experiment design, and optimization. The bottleneck shifts from generating ideas to validating them.

2

Publishing Crisis Forces Fundamental Restructuring

Discussed by: Nature, Science editors; peer review researchers; academic integrity watchdogs

The 50% productivity boost from LLMs, combined with paper mill fraud overwhelming peer review, triggers systemic failure. Retraction rates explode beyond 10,000 annually. Major journals implement mandatory AI detection and human verification protocols, creating multi-tier publication systems. Traditional peer review gives way to post-publication validation by specialized AI. Some fields abandon journals entirely for open preprint repositories with community verification. The transition is messy—prestigious labs see papers rejected despite solid science because reviewers can't distinguish from AI-polished junk. A two-year credibility crisis precedes new equilibrium.

3

AI Science Bubble Bursts After High-Profile Failures

Discussed by: MIT Technology Review, skeptical AI researchers, investors tracking 2025 hype correction

The $1.6 billion invested in AI science startups hits reality. Sakana's peer-reviewed paper is retracted after researchers discover it plagiarized established concepts. Periodic Labs and Lila Sciences struggle to translate computational predictions into real-world materials and drugs—the IBM Watson pattern repeats. OpenAI's 79x efficiency gain proves non-replicable outside controlled settings or requires human oversight that negates speed advantages. VCs who rushed in during 2025's gold rush write down valuations. The field doesn't die but resets expectations—AI becomes a useful tool rather than autonomous scientist, similar to how self-driving cars recalibrated from 'five years away' to gradual deployment.

4

Regulatory Intervention Fragments AI Science Development

Discussed by: Biosecurity researchers, EU AI Act implementers, academic integrity organizations

GPT-5's gene-editing optimization and autonomous lab systems trigger regulatory scrutiny. The EU, US, and China implement divergent rules: Europe restricts AI-designed biological experiments without oversight committees; China accelerates deployment in state labs; the US creates a patchwork of agency-specific regulations. Red Queen Bio's biosecurity focus becomes mandatory industry standard. Research bifurcates—computational discovery continues unrestricted, but autonomous wet-lab work faces lengthy approval processes. Brain drain accelerates toward permissive jurisdictions. By 2027, regulatory fragmentation creates scientific balkanization where breakthrough location depends more on legal framework than technical capability.

Historical Context

AlphaFold and the 2024 Nobel Prize

2020-2024

What Happened

DeepMind's AlphaFold solved the 50-year protein folding problem in 2020, predicting structures for 200 million proteins. The breakthrough earned Demis Hassabis and John Jumper the 2024 Nobel Prize in Chemistry—the first Nobel recognizing AI-enabled scientific discovery. Over 500,000 researchers adopted the open-source tool, generating thousands of papers on antibiotic resistance, drug discovery, and crop resilience.

Outcome

Short term: Proved AI could solve grand scientific challenges previously requiring decades of human effort; validated massive investment in AI for science.

Long term: Established template for LLMs tackling fundamental research problems; demonstrated that open-sourcing breakthrough AI accelerates global science rather than limiting it.

Why It's Relevant

AlphaFold's Nobel Prize validated AI's scientific capability one year before Science magazine recognized LLMs as 2025 breakthrough runners-up, creating momentum for the current AI science investment rush.

IBM Watson's Failed Drug Discovery Bet

2011-2019

What Happened

After Watson's 2011 Jeopardy triumph, IBM launched Watson for Drug Discovery, promising to revolutionize medicine by analyzing massive biomedical datasets. Major partnerships with Pfizer (2016) and MD Anderson followed. By 2019, IBM quietly discontinued the product after high-profile failures, including MD Anderson scrapping their installation. Watson made only small advances in drug discovery despite years of investment and publicity.

Outcome

Short term: Multi-million dollar partnerships dissolved; IBM shifted focus away from drug discovery toward clinical applications with more modest claims.

Long term: Created investor skepticism about AI drug discovery claims; established cautionary tale about over-promising AI capabilities before validation.

Why It's Relevant

Watson's trajectory warns that today's $1.6 billion in AI science startup funding could face similar reality checks when computational predictions must translate to real-world lab results.

The Paper Mill Crisis and Peer Review Collapse

2020-2025

What Happened

Paper retractions passed 10,000 in 2023, quadrupling over 20 years with the majority due to misconduct. In 2023, Hindawi retracted over 8,000 papers from organized paper mills selling fraudulent authorship. By 2025, AI tools like ChatGPT enabled industrial-scale generation of plagiarism-resistant junk science. Experts estimate one in 50 papers now show paper mill patterns, overwhelming traditional peer review gatekeeping.

Outcome

Short term: Major publishers implemented AI detection tools; peer review timelines lengthened; journal credibility declined in affected fields.

Long term: Forced fundamental questioning of peer review's viability; created opening for AI-assisted review systems despite concerns about AI reviewing AI-generated content.

Why It's Relevant

The existing publishing crisis means LLMs' 50% productivity boost arrives precisely when the system can least handle increased volume, potentially triggering complete restructuring of scientific validation.