A common but subtle user experience flaw in current AI models is their tendency to explain their process directly within the requested output. For example, a model might include a header like 'not trying to connect the ideas' instead of just performing the task. This meta-commentary breaks the illusion of a finished product and creates frustrating editing work.
An advanced workflow is emerging in OpenAI's Codex: the 'monothread.' Instead of fragmented chats, users maintain one continuous conversation. This leverages context compaction to build a long-term, evolving understanding of the user's projects, turning the AI into a persistent strategic partner for iterating on complex questions rather than a tool for one-off tasks.
OpenAI's GPT-5.5 is more expensive per token, but a new evaluation framework is emerging. The key metric isn't raw cost, but the model's efficiency in solving a problem. This 'intelligence per dollar' reframes cost analysis around performance and compute, where more expensive models can be cheaper overall if they solve tasks more efficiently.
While GPT-5.5 is a massive technical improvement, it may not feel transformative for 99% of users' daily workflows. Previous models like GPT-5.4 were already proficient enough for common tasks. The new model's value is realized at the ceiling of capability, on complex edge-case problems that stressed older models, rather than in everyday use.
A key breakthrough for GPT-5.5 is its stability in tasks running for over 7-8 hours, a feat previous models struggled with. This reliability is a game-changer for agentic AI, enabling complex software migrations and ambitious, long-running projects to execute autonomously without failing, fundamentally increasing the scope of work that can be delegated to AI.
OpenAI's GPT-5.5 launch featured a noticeable shift in communication towards humility and utility (e.g., 'We hope it's useful to you'). This contrasts sharply with competitor Anthropic's approach of hyping powerful models while withholding public access. The new strategy emphasizes iterative deployment and shipping, positioning OpenAI as pragmatic and user-focused.
Sophisticated users are moving beyond single-model setups. An optimal strategy involves using Anthropic's Opus 4.7 for its superior high-level planning capabilities and then handing off execution to OpenAI's GPT-5.5. This multi-model approach leverages the distinct strengths of each platform, widening the performance gap against any 'mono-model' workflow.
