After our first 20 issues and more than 100 AI × biology stories, a pattern is starting to emerge that no single paper, funding round, product launch, or benchmark can show on its own.
Yes, BAIO has covered a lot of news. But the point is that the news is beginning to add up. We’re starting to see a shape.
AI × biology is moving from models to systems. The bottlenecks are moving downstream. Virtual biology is climbing from proteins toward cells, patients, and physiology. Big AI companies are treating biology as a platform category.
And AI can produce impressive biological outputs. More important, though, is who validates it, who explains it, who has the right data, and who turns it into something scientists, clinicians, or companies can actually use.
So for this special report, I went back through BAIO’s first 20 issues and distilled eight signals that say where AI × biology stands in the first half of 2026.
Let’s dive in.
1. The field is moving from models to systems.
The early AI × biology story was often about models: better protein structure prediction, better molecular generation, better cell-state prediction, and so on. That still matters. But across BAIO’s first 20 issues, the more important shift is that models are becoming components inside larger systems.
A lot of work now is happening around the model: lab access, agent workflows, data pipelines, validation loops, deployment environments, regulatory traceability. A model connected to a wet lab, wrapped in a workflow a scientist can actually use, and able to learn from the next experiment is more useful than one that scores well in a notebook.
That is why the field increasingly looks less like a leaderboard race and more like an infrastructure race. The durable advantage may belong to whoever builds the tightest loop between prediction, experiment, feedback, and decision.
Further reading:
Issue 1: OpenAI connected GPT-5 to Ginkgo’s cloud lab for 36,000 closed-loop experiments; Phylo pitched an integrated biology environment.
Issue 3: Lilly and Insilico’s Prompt-to-Drug framework, plus Lilly’s $1 billion NVIDIA lab, pointed toward drug discovery as an industrial workflow.
Issue 5: Ginkgo opened its autonomous lab to outside researchers, turning lab automation into a platform.
Issue 13: Adaptyv Bio launched a wet-lab API for protein designers and AI agents.
Issue 16: Medra opened a 38,000-square-foot autonomous lab built around robots and closed-loop experimentation.
2. The bottleneck is moving downstream.
For years, the promise of AI in biology was that better prediction would make discovery faster. Now another pattern keeps showing up: once AI makes one step easier, the bottleneck moves to the next one.
If models can generate molecules, someone still has to make them. If agents can design experiments, someone still has to run them. If virtual cells can predict perturbations, someone still has to explain why the prediction is trustworthy. If AI can find more drug candidates, the hard part may shift to patient selection, clinical trial design, manufacturing, regulation, and development.
That is a useful corrective to both hype and dismissal.
Further reading:
Issue 3: Scienta Lab’s EVA targeted one of drug development’s nastiest bottlenecks - biology that works preclinically but fails to translate to patients.
Issue 7: LIBRIS showed the data-production bottleneck: AI could help optimize lipid nanoparticles, but first someone has to make enough formulations to train on.
Issue 9: X-Cell suggested that larger models alone are not enough; the gains came from pairing scale with diverse perturbation data.
Issue 10: JURA Bio attacked the build bottleneck directly by turning DNA synthesis itself into part of the generative process.
Issue 13: Warpspeed shifted attention from discovery to clinical trial design - patient selection, endpoints, and effect-size assumptions.
Issue 18: Formation Bio made the downstream argument explicit: the number of candidates has nearly doubled, FDA approvals have not.
Issue 20: 10x Science showed the characterization bottleneck: AI can generate protein drug candidates faster than the field can physically verify, measure, and explain them.
3. Virtual biology is climbing the stack.
In the world of virtual biology predicting protein structures or designing molecules remains foundational. But the ambition keeps moving upward: from molecules to proteins, from proteins to cells, from cells to tissues and organs, and now toward virtual patients.
A protein model asks: what does this molecule look like, bind, or do? A virtual cell asks: what happens if we perturb this gene, pathway, or program? A virtual patient asks something harder: how does biology unfold across time, disease, treatment, and clinical history?
This is still early, uneven, and easy to oversell, not least of all because the concept of a virtual cell mostly remains a vision. But one can see the direction. AI × biology is trying to build computational representations of living systems at increasing levels of abstraction. The closer those representations get to patients, the more useful - and more dangerous to trust too early - they become.
Further reading:
Issue 1: Isomorphic Labs and Terray showed the next step after AlphaFold - not just predicting structures, but predicting drug binding and finding druggable pockets.
Issue 7: A physics-based simulation followed a minimal bacterial cell from birth to division, while PerturbGen predicted how gene changes ripple through future cell states.
Issue 9: Xaira’s X-Cell pushed virtual cell models into unseen cell types and perturbations.
Issue 12: Meta’s TRIBE v2 modeled how the human brain responds to video, audio, and text - a step toward in-silico neuroscience.
Issue 13: Tripso decomposed cells into interpretable biological programs instead of one compressed state, surfacing biology that conventional analysis missed.
Issue 19: APOLLO compressed 33 years of hospital records from 7.2 million patients into virtual patient representations across 322 clinical tasks.
Issue 20: James Zou’s reported Human Intelligence raise put a name on a missing middle layer: a physiology foundation model, somewhere between protein models and virtual patients.

4. Big AI now treats biology as a platform category.
Google has Isomorphic Labs. OpenAI is funding life sciences through its Foundation and has already connected frontier models to biological experiments. Anthropic is building Claude into life sciences workflows, acquiring biology talent, and testing models on real biological design tasks. Meta is releasing brain models. In addition to selling GPUs to biotechs, NVIDIA is building models, infrastructure, datasets, and workflows for them. AWS is packaging AI-driven drug discovery as a service.
Biology is becoming one of the arenas where frontier AI companies compete for developers, enterprise customers, scientific legitimacy, and long-term platform control. The question is which AI company becomes the default operating layer for biological work.
Further reading:
Issue 1: Isomorphic Labs unveiled IsoDDE, a full drug design engine beyond AlphaFold, while OpenAI connected GPT-5 to Ginkgo’s lab infrastructure.
Issue 3: Anthropic’s Sonnet 4.6 nearly caught up with Opus on biology benchmarks, while Lilly and NVIDIA committed $1 billion to an AI drug discovery lab.
Issue 11: The OpenAI Foundation pledged at least $1 billion across life sciences, AI resilience, economic impact, and community programs, naming life sciences as its first priority.
Issue 12: Meta released TRIBE v2 for in-silico neuroscience, while Anthropic appeared to be building a dedicated biology mode for Claude.
Issue 14: Anthropic acquired Coefficient Bio, a small team of former Genentech AI drug discovery researchers, for more than $400 million.
Issue 16: Claude Mythos matched top human performers on Dyno Therapeutics’ biological design challenge, while NVIDIA pushed 1.7 million protein complex predictions into the AlphaFold Database.
Issue 17: Amazon Bio Discovery made the platform ambition explicit: AI agents, model design, lab partners, and validation workflows packaged for drug discovery.
5. Validation is the real dividing line.
AI × biology is full of impressive outputs: generated molecules, predicted structures, ranked targets, simulated cells, automated analyses, and fluent scientific reasoning.
No matter: you still need validation.
That can take different forms. Wet-lab validation. External cohorts. Prospective trials. FDA review. Randomized clinical data. Independent benchmarks. Sometimes even a failed prediction is useful, if it shows where the system breaks. The important question is “what kind of test did the claim pass?”
This is where the field becomes both more serious and more complicated. Some AI × biology systems are now producing experimentally validated designs, clinically useful predictions, and regulatory-grade evidence. Many others are still preprints, demos, internal benchmarks, or computational claims waiting for the lab.
Further reading:
Issue 5: Evo 2 moved from preprint to Nature with experimentally validated AI-designed DNA sequences, including bacteriophages that worked in living cells.
Issue 6: Perimeter’s Claire received FDA approval for use during breast cancer surgery after a pivotal trial showed improved margin assessment.
Issue 8: Origin Bio’s AI-designed DNA switches showed the opposite side of the line: impressive computational output, but no experimental validation yet and sharp criticism from domain experts.
Issue 9: Proteina-Complexa tested more than a million protein designs against 127 targets in a large wet-lab campaign.
Issue 10: The Nature Biotechnology GBAI review made the field’s validation problem explicit: many biological foundation models still lack robust benchmarks and wet-lab proof.
Issue 11: A pathology AI model estimated Oncotype DX recurrence scores from routine slides and was validated against randomized trial data in The Lancet Oncology.
Issue 12: Latent-Y produced lab-confirmed antibody binders from text prompts, while also showing why human review still matters when agents make biologically impossible design choices.
Issue 16: CompBioBench gave AI agents 100 computational biology tasks with verifiable answers, exposing both strong performance and brittle failure modes.

In mainstream AI, scale often means more parameters, more compute, more tokens. In biology, that story is incomplete. The recurring lesson is that better models need better biological coverage: more cell types, more perturbations, more species, more patient cohorts, more experimental conditions, more modalities.
That matters because biology is not one distribution. A model trained on familiar proteins, common cell lines, clean benchmark datasets, or narrow patient populations may look strong until it meets a new tissue, organism, disease context, or clinical setting. The field needs datasets that expose models to the messy variation of living systems.
This may become one of the deepest moats in AI × biology. Compute can be bought. Model architectures spread. But causally rich, diverse, well-annotated biological data - especially linked to experiments and outcomes - is much harder to build.
Further reading:
Issue 1: Arc Institute cut perturb-seq costs and partnered with Tahoe AI and Biohub to build a 120-million-cell open dataset for virtual cell models.
Issue 9: Xaira’s X-Cell showed that scaling parameters was not enough; gains came from 152,000 unique experimental conditions and diverse perturbation data.
Issue 10: Basecamp Research launched the Trillion Gene Atlas to expand known genetic diversity 100-fold, arguing that biological AI scales better with diverse data than compute alone.
Issue 11: Cerebra was validated across 3 million patients from four healthcare systems, showing why clinical AI needs to survive fragmented, incomplete, real-world data.
Issue 15: Renaissance Philanthropy funded ten projects targeting missing datasets for AI biology, from protein binding affinity to bioimaging ground truth and 20,000-species sequencing.
Issue 19: APOLLO drew power from breadth: 33 years of hospital records, 7.2 million patients, and 160 million clinical events compressed into virtual patient representations.
7. Access is moving faster than trust.
One of the strangest shifts in AI × biology is that the barrier to trying things is falling faster than the barrier to knowing whether they worked.
That does not only mean better tools for professionals. It means a non-biologist can use public AI tools to help design a dog cancer vaccine. Someone with no lab training can sequence their own genome at home with a MinION, Claude-written protocols, and a camping table. A small team can use a browser-based protein platform to redesign the binding loops of Keytruda in minutes.
This is not the same as saying the results are reliable. Often they are not proven, not peer-reviewed, not clinically useful, or not validated in the lab. But that is sort of the point. AI is making biological capability more accessible before it has made biological judgment more scalable.
Further reading:
Issue 8: A non-biologist used ChatGPT and AlphaFold to help design a personalized cancer vaccine for his dog - fascinating, but N=1, uncontrolled, and not proof of efficacy.
Issue 8: Origin Bio released 10,000 AI-designed DNA switches before experimental validation, drawing sharp criticism from Stanford’s Anshul Kundaje.
Issue 20: “Vibe genomics” showed someone with no lab experience sequencing their own genome at home with Claude-written protocols and consumer-accessible hardware.
Issue 20: LiteFold put protein redesign into a browser, using a plain-English interface to generate redesigned Keytruda variants - computational only, but a striking signal about falling barriers.
8. Explainability is becoming infrastructure.
As AI moves deeper into biology, “the model says so” becomes less useful. A drug target, protein function, clinical prediction, or experimental recommendation is only valuable if scientists can inspect why the system reached it - and decide whether the reasoning holds up.
Explainability matters for debugging models, designing follow-up experiments, convincing collaborators, satisfying regulators, and knowing when a system is confidently wrong. In biology, trust is built by showing the evidence chain, not by nicer chat interfaces.
Further reading:
Issue 10: BioReason-Pro predicted protein function while showing step-by-step reasoning, with experts preferring its annotations over curated UniProt entries in blinded evaluation.
Issue 13: Tripso decomposed cells into separate biological programs rather than one compressed state, making it easier to see which pathways or transcription-factor programs drove a result.
Issue 16: CompBioBench exposed why agents fail: not only from lack of knowledge, but from stopping at plausible-looking analyses before resolving ambiguity.
Issue 18: VCR-Agent generated biological hypotheses with explicit reasoning traces that scientists could inspect, rather than just giving an answer.
Issue 20: 10x Science made explainability central to molecular characterization, showing which fragments matched, at what coverage, and with what error.

