top of page

Biology x AI

  • Foto van schrijver: Tung
    Tung
  • 26 mrt
  • 5 minuten om te lezen

Biology is undergoing a structural transition from lab experimentation to one increasingly shaped by data, compute, and simulation. Discovery is moving upstream, from physical experimentation toward digital modeling. Instead of testing thousands of compounds in the lab, researchers can now simulate outcomes, narrow probabilities, and only validate the highest-conviction paths physically. The implication is profound: biology is becoming computable. As with every previous paradigm shift, the key question is not which application wins, but where value accrues in the stack.


Data Layer: The Oil of Biology

At the foundation of this shift lies biological data. Genomics, proteomics, and patient-level datasets form one of the largest and fastest-growing data domains globally, embedded within a healthcare system worth more than $10 trillion. Yet the value is not in raw data itself. Biological data is fragmented, unstructured, and often siloed across institutions, making it difficult to use effectively.


The real opportunity lies in aggregation and standardization. Systems that turn fragmented datasets into structured, model-ready inputs create compounding advantages. As more data is added, models improve; as models improve, more users join, and as more users join, more data is generated. This creates a feedback loop that strengthens over time. Fields like gene editing and longevity are particularly data-hungry. Understanding gene expression, aging pathways, or peptide interactions requires vast, high-quality datasets. However, from an investment perspective, data alone rarely captures the majority of value. True value often concentrates elsewhere in the stack, moving beyond raw information toward integrated infrastructure and compounding moats. These are systems that become more difficult to displace as they scale.


Compute + Simulation Layer As biology becomes more data-driven, compute emerges as the central bottleneck. Molecular simulations, protein folding, and drug discovery workflows increasingly rely on large-scale computation rather than purely physical experimentation. This transition directly challenges Eroom's Law, which has historically seen drug development become slower and more expensive over time.


Instead of iterating blindly in the lab, researchers can simulate outcomes, dramatically narrowing the search space. This applies across domains:

-Gene editing: predicting CRISPR outcomes and off-target effects

-Synthetic biology: designing biological circuits

-Peptides: optimizing structure and function

-Longevity: modeling aging pathways


The economic implications are significant. The pharmaceutical industry alone represents a market of roughly $1.5 trillion, with tens of billions spent annually on research and development. Even marginal improvements in efficiency can unlock enormous value. But more importantly, every participant in the ecosystem, from biotech startups to large pharmaceutical companies, depends on compute. This makes it a horizontal layer with structural demand. Unlike single-drug companies, compute providers are not exposed to binary outcomes. Their revenues scale with usage, making this one of the highest potential return layers in the entire stack.


Robotics Layer While compute enables simulation, biology ultimately still interacts with the physical world. This is where robotics becomes essential. If compute is the brain of the system, robotics is the hands. Automated laboratories, high-throughput experimentation systems, and robotic workflows allow biological hypotheses to be tested rapidly and at scale.


Robotics closes the loop between digital and physical biology. AI models generate predictions, robotic systems validate them, and the resulting data feeds back into the models. This creates a continuous cycle of improvement: simulate, test, learn, and iterate.

This is particularly relevant in areas like synthetic biology and peptide discovery, where rapid iteration is critical. Instead of slow, manual experimentation, robotic systems enable parallel testing at scale. Over time, this transforms biology into a continuous, self-improving system.


From an economic perspective, robotics introduces hardware complexity and capital intensity. However, it also creates strong integration and switching costs. As a result, it acts as a powerful accelerator for the entire stack, with high, but slightly lower, ROI potential compared to pure compute and platform layers.


Privacy / Secure Data Layer A defining feature of biology is the sensitivity of its data. Medical records, genomic data, and patient histories are among the most regulated datasets in existence. At the same time, they are essential for advancing areas like longevity research and personalized medicine. This creates a structural constraint. Without access to high-quality data, models cannot improve; yet unrestricted access is not possible. The solution lies in privacy-preserving infrastructure, enabling data to be used without being exposed. Technologies such as federated learning and encrypted computation allow institutions to collaborate without sharing raw data. This is particularly important for global datasets in aging research or large-scale clinical studies.


While this layer is critical, it operates on a longer timeline. Adoption is slower due to regulatory and technical complexity. However, once established, it becomes deeply embedded and difficult to replace, making it a high-upside but delayed-return layer.


AI Protein and Molecular Modeling At the core of computational biology lies protein and molecular modeling. Breakthroughs such as AlphaFold have demonstrated that predicting protein structures can be solved computationally. Proteins underpin nearly all biological processes, making this layer foundational. This capability extends across multiple domains:

-Drug discovery

-Peptide engineering

-Synthetic biology

-Longevity pathways


However, from an investment perspective, this layer is less defensible. Models improve rapidly and are often open-sourced, reducing long-term differentiation. As a result, while essential, this layer tends to see value shift toward infrastructure and platforms that integrate and operationalize these models.


Bio-AI Platforms As the stack matures, the individual layers converge into integrated platforms. Researchers do not want to manage separate systems for data, compute, robotics, and modeling. They want unified environments that abstract complexity into usable workflows. This creates the opportunity for full-stack platforms that combine all layers into a single system. These platforms sit at the center of the ecosystem, connecting data, compute, experimentation, and application. They are particularly powerful in enabling emerging fields:

-Gene editing platforms that integrate prediction and validation

-Synthetic biology toolchains for designing biological systems

-Peptide discovery environments

-Longevity research platforms combining full-stack biological data

Because these platforms are embedded in workflows, they benefit from high switching costs and recurring revenue. They capture value across multiple layers, making them one of the most attractive areas for long-term returns.


Simulation Engines

The ultimate destination of this stack is full biological simulation. The goal is not just to predict individual molecules, but modeling entire systems: cells, pathways, and eventually organisms. This is particularly relevant for longevity and age reversal. Aging is not a single pathway but a complex system of interacting processes. Simulation enables researchers to model these interactions rather than targeting isolated mechanisms.

Similarly, synthetic biology aims to design entirely new biological systems. This requires the ability to simulate complex interactions before physical implementation.


If successful, this layer transforms biology into a programmable system. Experiments can be conducted in simulation, dramatically reducing cost and time. While still early, this represents the highest long-term upside in the stack.


Traditional Biotech Companies vs. Platform Dominance

Despite these structural changes, much of the investment landscape remains focused on traditional biotech companies, particularly those centered on single-drug pipelines.

This is especially visible in areas like peptides and longevity, where hype cycles often drive capital into companies dependent on individual outcomes. These businesses are inherently fragile, with binary risk profiles and long timelines. They do not compound in the same way as infrastructure or platform layers. As a result, they represent one of the least attractive areas from a risk-adjusted return perspective.


The highest-value opportunities lie in layers that serve the entire ecosystem. Compute infrastructure, integrated platforms, and enabling technologies such as robotics and privacy systems all share key characteristics: they scale with usage, generate recurring revenue, and benefit from network effects. These layers capture value from every successful application, whether in gene editing, synthetic biology, peptides, or longevity, rather than relying on any single outcome. This creates compounding growth and durable competitive advantages over time.


Final Thought

The future of biology will not be defined by a single breakthrough drug, therapy, or technology. It will be defined by the systems that make continuous discovery possible across all domains. As biology becomes increasingly computable, the layers that enable this transformation: compute, platforms, and integrated infrastructure will capture the majority of value. The opportunity is not in predicting which applications win, but in identifying the systems that make all applications possible.







Opmerkingen


Het is niet meer mogelijk om opmerkingen te plaatsen bij deze post. Neem contact op met de website-eigenaar voor meer info.
bottom of page