Instead of chasing the latest hyped AI model, focus on building modular, system-based workflows. This allows you to easily plug in new, better models as they are released, instantly upgrading your capabilities without having to start over.
Overly structured, workflow-based systems that work with today's models will become bottlenecks tomorrow. Engineers must be prepared to shed abstractions and rebuild simpler, more general systems to capture the gains from exponentially improving models.
Simply offering the latest model is no longer a competitive advantage. True value is created in the system built around the model—the system prompts, tools, and overall scaffolding. This 'harness' is what optimizes a model's performance for specific tasks and delivers a superior user experience.
Building an AI-native product requires betting on the trajectory of model improvement, much like developers once bet on Moore's Law. Instead of designing for today's LLM constraints, assume rapid progress and build for the capabilities that will exist tomorrow. This prevents creating an application that is quickly outdated.
In the fast-paced world of AI, focusing only on the limitations of current models is a failing strategy. GitHub's CPO advises product teams to design for the future capabilities they anticipate. This ensures that when a more powerful model drops, the product experience can be rapidly upgraded to its full potential.
When developing AI-powered tools, don't be constrained by current model limitations. Given the exponential improvement curve, design your product for the capabilities you anticipate models will have in six months. This ensures your product is perfectly timed to shine when the underlying tech catches up.
To avoid being made obsolete by the next foundation model (e.g., GPT-5), entrepreneurs must build products that anticipate model evolution. This involves creating strategic "scaffolding" (unique workflows and integrations) or combining LLMs with proprietary data, like knowledge graphs, to create a defensible business.
Exceptional AI content comes not from mastering one tool, but from orchestrating a workflow of specialized models for research, image generation, voice synthesis, and video creation. AI agent platforms automate this complex process, yielding results far beyond what a single tool can achieve.
The pace of AI model improvement is faster than the ability to ship specific tools. By creating lower-level, generalizable tools, developers build a system that automatically becomes more powerful and adaptable as the underlying AI gets smarter, without requiring re-engineering.
Instead of building monolithic agents, create modular sub-workflows that function as reusable 'tools' (e.g., an 'image-to-video' tool). These can be plugged into any number of different agents. This software engineering principle of modularity dramatically speeds up development and increases scalability across your automation ecosystem.
To maximize AI's impact, don't just find isolated use cases for content or demand gen teams. Instead, map a core process like a campaign workflow and apply AI to augment each stage, from strategy and creation to localization and measurement. AI is workflow-native, not function-native.