Fine-tuning creates model-specific optimizations that quickly become obsolete. Blitzy favors developing sophisticated, system-level "memory" that captures enterprise-specific context and preferences. This approach is model-agnostic and more durable as base models improve, unlike fine-tuning which requires constant rework.
Effective enterprise AI needs a contextual layer—an 'InstaBrain'—that codifies tribal knowledge. Critically, this memory must be editable, allowing the system to prune old context and prioritize new directives, just as a human team would shift focus from revenue growth one quarter to margin protection the next.
The next major evolution in AI will be models that are personalized for specific users or companies and update their knowledge daily from interactions. This contrasts with current monolithic models like ChatGPT, which are static and must store irrelevant information for every user.
Simply offering the latest model is no longer a competitive advantage. True value is created in the system built around the model—the system prompts, tools, and overall scaffolding. This 'harness' is what optimizes a model's performance for specific tasks and delivers a superior user experience.
The key for enterprises isn't integrating general AI like ChatGPT but creating "proprietary intelligence." This involves fine-tuning smaller, custom models on their unique internal data and workflows, creating a competitive moat that off-the-shelf solutions cannot replicate.
Early-stage AI startups should resist spending heavily on fine-tuning foundational models. With base models improving so rapidly, the defensible value lies in building the application layer, workflow integrations, and enterprise-grade software that makes the AI useful, allowing the startup to ride the wave of general model improvement.
The "agentic revolution" will be powered by small, specialized models. Businesses and public sector agencies don't need a cloud-based AI that can do 1,000 tasks; they need an on-premise model fine-tuned for 10-20 specific use cases, driven by cost, privacy, and control requirements.
Basic supervised fine-tuning (SFT) only adjusts a model's style. The real unlock for enterprises is reinforcement fine-tuning (RFT), which leverages proprietary datasets to create state-of-the-art models for specific, high-value tasks, moving beyond mere 'tone improvements.'
Enterprises will shift from relying on a single large language model to using orchestration platforms. These platforms will allow them to 'hot swap' various models—including smaller, specialized ones—for different tasks within a single system, optimizing for performance, cost, and use case without being locked into one provider.
The founder of Stormy AI focuses on building a company that benefits from, rather than competes with, improving foundation models. He avoids over-optimizing for current model limitations, ensuring his business becomes stronger, not obsolete, with every new release like GPT-5. This strategy is key to building a durable AI company.
Fine-tuning remains relevant but is not the primary path for most enterprise use cases. It's a specialized tool for situations with unique data unseen by foundation models or when strict cost and throughput requirements for a high-volume task justify the investment. Most should start with RAG.