The test intentionally used a simple, conversational prompt one might give a colleague ("our blog is not good...make it better"). The models' varying success reveals that a key differentiator is the ability to interpret high-level intent and independently research best practices, rather than requiring meticulously detailed instructions.

Related Insights

People struggle with AI prompts because the model lacks background on their goals and progress. The solution is 'Context Engineering': creating an environment where the AI continuously accumulates user-specific information, materials, and intent, reducing the need for constant prompt tweaking.

With models like Gemini 3, the key skill is shifting from crafting hyper-specific, constrained prompts to making ambitious, multi-faceted requests. Users trained on older models tend to pare down their asks, but the latest AIs are 'pent up with creative capability' and yield better results from bigger challenges.

Before delegating a complex task, use a simple prompt to have a context-aware system generate a more detailed and effective prompt. This "prompt-for-a-prompt" workflow adds necessary detail and structure, significantly improving the agent's success rate and saving rework.

Current AI models often provide long-winded, overly nuanced answers, a stark contrast to the confident brevity of human experts. This stylistic difference, not factual accuracy, is now the easiest way to distinguish AI from a human in conversation, suggesting a new dimension to the Turing test focused on communication style.

To get the best results from AI, treat it like a virtual assistant you can have a dialogue with. Instead of focusing on the perfect single prompt, provide rich context about your goals and then engage in a back-and-forth conversation. This collaborative approach yields more nuanced and useful outputs.

The early focus on crafting the perfect prompt is obsolete. Sophisticated AI interaction is now about 'context engineering': architecting the entire environment by providing models with the right tools, data, and retrieval mechanisms to guide their reasoning process effectively.

Getting a useful result from AI is a dialogue, not a single command. An initial prompt often yields an unusable output. Success requires analyzing the failure and providing a more specific, refined prompt, much like giving an employee clearer instructions to get the desired outcome.

AI development has evolved to where models can be directed using human-like language. Instead of complex prompt engineering or fine-tuning, developers can provide instructions, documentation, and context in plain English to guide the AI's behavior, democratizing access to sophisticated outcomes.

Good Star Labs found GPT-5's performance in their Diplomacy game skyrocketed with optimized prompts, moving it from the bottom to the top. This shows a model's inherent capability can be masked or revealed by its prompt, making "best model" a context-dependent title rather than an absolute one.

Asking an AI to 'predict' or 'evaluate' for a large sample size (e.g., 100,000 users) fundamentally changes its function. The AI automatically switches from generating generic creative options to providing a statistical simulation. This forces it to go deeper in its research and thinking, yielding more accurate and effective outputs.