We scan new podcasts and send you the top 5 insights daily.
Unlearn.ai strategically avoids diseases where a single biomarker determines progression. Instead, they focus on complex, systematic diseases where many variables each have a small impact on the outcome. These are the areas where sophisticated, multi-variable modeling provides the most significant advantage over standard statistical adjustment.
A significant part of Unlearn.ai's value is not just its advanced generative models, but its painstaking data harmonization work. The company builds internal machine learning tools to unify complex, disparate data sources like clinical trials and real-world data, which is the essential foundation for creating powerful models.
By using foundation models to analyze vast datasets, companies can create a synthetic 'standard of care' arm for single-arm Phase 1 trials. The AI matches patients based on deep clinical and genomic parameters, providing insights comparable to a much larger Phase 3 trial.
Unlearn.ai's method for late-phase trials (PROCOVA) is acceptable to regulators because it's designed to statistically correct for any bias in the digital twin model. This ensures the model's inaccuracy doesn't affect the trial's final decision procedure or error rate, a critical feature distinguishing it from simply replacing the control arm.
Broad diagnostic categories like 'diabetes' or 'insomnia' likely encompass several distinct underlying conditions. Continuous data streams from wearables and CGMs can help researchers identify these subtypes, paving the way for more personalized treatments.
Instead of the high-risk approach of replacing a trial's control arm with digital twins, Unlearn.ai adds counterfactual data to every participant. This method increases a trial's statistical power, allowing for smaller control arms or a higher chance of success, while satisfying regulatory constraints for pivotal trials.
While AI excels where large, clean datasets exist (like protein folding), it struggles with modeling slow, progressive diseases like Alzheimer's or obesity. These are organ-level phenomena, and the necessary data doesn't exist yet. In vivo platforms are critical for generating this required foundational data.
By feeding an AI agent diverse personal data—diet logs, sleep tracking, bloodwork, and genetics—it can identify complex health issues that elude general advice. The AI can find "needle in the haystack" answers, like connecting restless leg syndrome to Swedish ancestry, offering hyper-personalized insights.
It's impossible to generate human data at the scale of in silico experiments. The key is to create highly accurate simulations of human physiology (digital twins) and then validate their predictions with limited, strategic human data. If the model proves reliable, it could drastically accelerate R&D.
The low-hanging fruit of finding a single predictive biomarker is gone. The next frontier for bioinformatics is developing complex, 'multimodal models' that integrate several data points to predict outcomes. The key challenge is creating sophisticated models that still yield practical, broadly applicable clinical insights.
Unlearn.ai found that scaling digital twins from CNS to oncology isn't about parameter changes. Radically different data structures—like oncology's hierarchy of rare diseases and complex treatment histories—demand entirely new modeling approaches, unlike the more siloed data found in CNS trials.