Get your free personalized podcast brief

We scan new podcasts and send you the top 5 insights daily.

Jurgi Camblong cautions against the hype that Large Language Models (LLMs) can solve every problem in medicine. Sophia Genetics uses a diverse "toolbox" of AI—including statistical methods and machine learning—selecting the most efficient mathematical model for a specific biological problem and dataset.

Related Insights

Unlike image recognition or NLP, clinical trial data possesses a unique and complex mathematical geometry. According to Dr. Juraji, this means generic AI models are insufficient. Solving trial failures requires specialized AI built to navigate this specific, difficult data landscape.

Powerful AI models for biology exist, but the industry lacks a breakthrough user interface—a "ChatGPT for science"—that makes them accessible, trustworthy, and integrated into wet lab scientists' workflows. This adoption and translation problem is the biggest hurdle, not the raw capability of the AI models themselves.

In high-stakes fields like pharma, AI's ability to generate more ideas (e.g., drug targets) is less valuable than its ability to aid in decision-making. Physical constraints on experimentation mean you can't test everything. The real need is for tools that help humans evaluate, prioritize, and gain conviction on a few key bets.

Resist the urge to apply LLMs to every problem. A better approach is using a 'first principles' decision tree. Evaluate if the task can be solved more simply with data visualization or traditional machine learning before defaulting to a complex, probabilistic, and often overkill GenAI solution.

Dr. Juraji argues against a single "do-it-all" AI. Instead, he envisions a future of "speciated" AI systems where different modules, like the lobes of a brain (e.g., LLMs, causal AI), work together to tackle the multifaceted challenges of drug development.

Early AI drug discovery platforms built robust models but often failed to generate relevant outputs. Their lack of deep biological understanding led to flawed data collection and training sets, creating a "garbage in, garbage out" problem where models were disconnected from real-world biology.

A major misconception is that general-purpose Large Language Models (LLMs) can be readily applied to complex biological problems. Biological data, like RNA sequencing, constitutes a unique language that requires custom-built foundation models, not simply fine-tuning of existing LLMs.

Generative AI is not viewed as a standalone solution for drug discovery. Alloy's perspective is that its primary value is in enhancing and automating existing workflows. The model requires a 'lab in the loop' and 'human in the loop,' where AI assists scientists by making them more efficient and improving data analysis, rather than replacing the core wet lab process.

While acknowledging the power of Large Language Models (LLMs) for linear biological data like protein sequences, CZI's strategy recognizes that biological processes are highly multidimensional and non-linear. The organization is focused on developing new types of AI that can accurately model this complexity, moving beyond the one-dimensional, sequential nature of language-based models.

Achieving explainability in AI for drug development isn't about post-hoc analysis. It requires building models from the ground up using inherently interpretable data like RNA sequencing and mutational profiles. When the inputs are explainable, the model's outputs become explainable by design.