We scan new podcasts and send you the top 5 insights daily.
Ceramic AI founder Anna Patterson explains their pivot from training to search was driven by a key insight: providing models with live data via low-cost search is far more efficient and timely than the expensive, slow process of continuous retraining.
Venture capitalists are increasingly being pitched by AI startups claiming to have solved "continual learning." However, many of these are simply using clever workarounds, like giving a model a 'scratch pad' to reference new data, rather than building models that can fundamentally learn and update themselves in real-time.
According to Anna Patterson, vector databases struggle with scale, as distinguishing between billions of items requires increasingly long vectors. Their "soft match" functionality also creates relevancy challenges, forcing enterprises to become search experts to tune results, unlike more traditional keyword-based systems.
AI's hunger for context is making search a critical but expensive component. As illustrated by Turbo Puffer's origin, a single recommendation feature using vector embeddings can cost tens of thousands per month, forcing companies to find cheaper solutions to make AI features economically viable at scale.
Instead of expensive, static pre-training on proprietary data, enterprises prefer RAG. This approach is cheaper, allows for easy updates as data changes, and benefits from continuous improvements in foundation models, making it a more practical and dynamic solution.
While early AI development requires constant testing of new models, Conative.ai found they eventually reached a stable architecture. The focus then shifted from wholesale model replacement to fine-tuning existing layers with specific data, reducing the pressure to chase every new innovation.
By training a smaller, specialized model where company data is in the weights, firms avoid the high token costs of repeatedly feeding context to large frontier models. This makes complex, data-intensive workflows significantly cheaper and faster.
Ceramic AI's drastically lower search cost (5 cents/1k queries) enables novel AI workflows. For example, "supervised generation" involves an AI continuously fact-checking its own output via multiple, inexpensive searches, building trust in high-stakes applications like legal brief writing.
Yahoo built its AI search engine, Scout, not by training a massive model, but by using a smaller, affordable LLM (Anthropic's Haiku) as a processing layer. The real power comes from feeding this model Yahoo's 30 years of proprietary search data and knowledge graphs.
Unlike chatbots that rely solely on their training data, Google's AI acts as a live researcher. For a single user query, the model executes a 'query fanout'—running multiple, targeted background searches to gather, synthesize, and cite fresh information from across the web in real-time.
The key to a truly intelligent enterprise AI is not a static model, but one that uses reinforcement learning (RL) to continuously update its own weights overnight based on daily interactions, a concept known as 'continuous learning'.