AI is reshaping scientific discovery faster than most organizations can absorb it. Hypothesis generation, analysis, prediction, the pace of all of it is accelerating in ways that would have been difficult to imagine just a few years ago. And yet, for all the energy and investment pouring into AI for life sciences and broader R&D, something fundamental has been missing from the conversation.
It isn't the algorithms. It isn't the data volume. It's the structure.
Here's the thing I think matters most right now: the more structured the underlying data systems are, the more effective AI is going to be. That's not a nuanced qualifier. It's the defining challenge in front of our industry, and it's one that most organizations are still underestimating.
Scientific Data Has Been Digitized. It Has Not Been Orchestrated.
Over the past two decades, laboratories have made extraordinary investments in digital tools, instruments that generate vast amounts of data, systems that capture experimental records and platforms that enable collaboration. The digitization of science has been real and meaningful.
But digitization is not orchestration. And that distinction matters more now than it ever has.
Most scientific work today remains fragmented across tools, teams, and stages of the R&D lifecycle. Experimental records are captured, but the continuity of intent, process, material state, and decision-making is routinely lost as projects move from discovery into development and eventually into manufacturing. Critical context, the kind that makes results reproducible and insights actionable, doesn't travel with the work. It stays behind.
This is not a technology failure. It's a representation failure. Science has rarely been defined in a way that makes it reproducible, scalable, or operationally consistent, not because people weren't trying, but because the tools and frameworks to do it simply didn't exist.
Until now.
What Is Scientific Process Modeling — and Why Does AI Need It?
Legacy electronic lab notebook systems were built around a fundamentally narrative model of scientific record-keeping. A scientist does an experiment, writes it up, and the system captures the story. That approach served a real purpose, but it is not a foundation for the kind of structured, AI-ready science that modern R&D demands.
What's required is a process modeling approach, one where every step of an experiment, whether it's wet lab or dry lab, computational or physical, is precisely defined. Inputs into containers, outputs out of containers, transformations, decisions, conditions, all of it modeled explicitly rather than described loosely. Not as narrative, but as structure.
Before AI, I think it would have been too difficult to ask scientists to set up their experiments with that level of detail. The overhead would have exceeded the benefit. But AI changes that equation entirely. It provides the means to configure experimental workflows at a level of fidelity that wasn't previously practical, and in doing so, it creates a scaffold, a structured backbone, to which every piece of data throughout a scientific project can be attached.
This is what has been missing from R&D software since its inception: an effective, continuous model of the process being tracked. Not a record of what happened, but a living representation of how it happened, and why.
Data Lineage in R&D: How AI Traces Every Experiment and Decision
The concept of lineage is central to why this matters, and it's one that I think the industry hasn't fully grappled with yet.
When scientific processes are explicitly modeled, you gain the ability to connect every piece of information to every other piece of information. What material was created from what precursor. What process step produced what output. What conditions governed what transformation. All of it laid out, not just in documentation, but in the underlying data structure of the system itself.
That lineage is what gives AI a stable, contextualized foundation to work from. AI cannot reliably predict, optimize, or guide scientific work without understanding how that work is actually performed. It can analyze data, but it cannot reason meaningfully about processes that are not explicitly defined. When experimental context, material state, and decision history are modeled together, AI can move from retrospective analysis to forward-looking reasoning. Models become explainable. Scientific decision-making can be accelerated, not just recorded.
This also means that precision in how materials themselves are represented matters enormously. The degree of fidelity with which biological components, from small molecules to complex biologics, are captured in a system determines how much that system can do with the data it holds. Higher fidelity representation leads directly to more powerful capabilities, capabilities that simply aren't available to systems working from lower-resolution data.
A Common Data Language from Discovery to Manufacturing
Perhaps the most significant, and most under appreciated, implication of this approach is the possibility of a common language across research, development, manufacturing, and automation.
Historically, these have been distinct worlds with distinct vocabularies, distinct systems, and distinct teams. The handoffs between them have been a persistent source of inefficiency, risk, and lost knowledge. Scientific insight generated in discovery doesn't consistently travel with the process knowledge, material lineage, and decision rationale needed to make it actionable downstream.
One of the greatest untapped opportunities in life sciences is addressing exactly this gap. When the same process modeling framework governs how work is represented across all stages of the lifecycle, it becomes possible to move back and forth between those stages in a genuinely seamless way. Human-executed steps and machine-executed steps can coexist within the same workflow, defined in the same language, connected to the same data model.
This is not a distant aspiration. It is the structural foundation that makes the vision of end-to-end continuity from molecule to market real and achievable.
How Dotmatics and Siemens Create an End-to-End R&D Digital Thread
Siemens has spent decades building leadership in digital twins and lifecycle management across engineering and manufacturing, connecting complex processes in industries where precision, traceability, and scale are non-negotiable. That expertise has created an extraordinary foundation for managing the operational complexity of the downstream world.
What has been missing is the extension of that foundation upstream, into the earliest stages of discovery and research, where the science originates, where the decisions that determine development trajectories are made, and where the data that should inform everything downstream is first created.
That's what Dotmatics brings. And together, the combination enables something genuinely new: a connected digital continuum that spans the entire innovation lifecycle, from the design of a molecule through its development, scale-up, and production.
This isn't about layering a discovery tool onto a manufacturing platform. It's about establishing structural continuity across the full R&D value chain, so that scientific work conducted at the bench informs and connects to everything that follows. The digital thread extends in both directions, upstream into discovery and downstream into development and manufacturing, creating a unified, coordinated system where knowledge doesn't get lost at the handoffs.
Building the Structured Foundation AI Needs to Deliver in Life Sciences
AI is going to keep advancing. The pace of hypothesis generation, analysis, and prediction will continue to accelerate. But the organizations that capture the real value from AI in R&D will not simply be the ones with access to the best models. They will be the ones that have built the structural foundation, the process models, the material lineage, the workflow continuity, that allows AI to operate reliably, contextually, and at scale.
Structured science is not a constraint on innovation. It is the multiplier of it.
The most exciting thing about this moment, and I genuinely believe this, is that the tools to do this right are finally available. The ability to model scientific work with the fidelity required to make AI meaningful, to connect discovery to development to manufacturing in a single coherent system, to preserve the context that has historically been lost at every transition, that ability is real today in a way it has never been before.
The question for every life sciences organization is whether they're building on a foundation designed for this era, or one designed for the last.
