New AI tools often have flawed user experiences. Instead of just getting frustrated, create a detailed product breakdown with recommendations for improvement. Sending this to the company serves as a powerful "warm intro," showcasing your product skills and providing value before you're hired.

Related Insights

Don't treat evals as a mere checklist. Instead, use them as a creative tool to discover opportunities. A well-designed eval can reveal that a product is underperforming for a specific user segment, pointing directly to areas for high-impact improvement that a simple "vibe check" would miss.

Many teams wrongly focus on the latest models and frameworks. True improvement comes from classic product development: talking to users, preparing better data, optimizing workflows, and writing better prompts.

After testing a prototype, don't just manually synthesize feedback. Feed recorded user interview transcripts back into the original ChatGPT project. Ask it to summarize problems, validate solutions, and identify gaps. This transforms the AI from a generic tool into an educated partner with deep project context for the next iteration.

People struggle with AI prompts because the model lacks background on their goals and progress. The solution is 'Context Engineering': creating an environment where the AI continuously accumulates user-specific information, materials, and intent, reducing the need for constant prompt tweaking.

In AI PM interviews, 'vibe coding' isn't a technical test. Interviewers evaluate your product thinking through how you structure prompts, the user insights you bring to iterations, and your ability to define feedback loops, not your ability to write code.

The best agentic UX isn't a generic chat overlay. Instead, identify where users struggle with complex inputs like formulas or code. Replace these friction points with a native, natural language interface that directly integrates the AI into the core product workflow, making it feel seamless and powerful.

Developers often test AI systems with well-formed, correctly spelled questions. However, real users submit vague, typo-ridden, and ambiguous prompts. Directly analyzing these raw logs is the most crucial first step to understanding how your product fails in the real world and where to focus quality improvements.

Open-ended prompts overwhelm new users who don't know what's possible. A better approach is to productize AI into specific features. Use familiar UI like sliders and dropdowns to gather user intent, which then constructs a complex prompt behind the scenes, making powerful AI accessible without requiring prompt engineering skills.

The most effective AI user experiences are skeuomorphic, emulating real-world human interactions. Design an AI onboarding process like you would hire a personal assistant: start with small tasks, verify their work to build trust, and then grant more autonomy and context over time.

Traditional hiring assessments that ban modern tools are obsolete. A better approach is to give candidates access to AI tools and ask them to complete a complex task in an hour. This tests their ability to leverage technology for productivity, not their ability to memorize information.